[ 464.312914] env[61006]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61006) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.313372] env[61006]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61006) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.313372] env[61006]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61006) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.313666] env[61006]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 464.407340] env[61006]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61006) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 464.417356] env[61006]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61006) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 465.016790] env[61006]: INFO nova.virt.driver [None req-826a0a57-3f60-42da-b715-c3851c4c71bb None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 465.087073] env[61006]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.087252] env[61006]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.087350] env[61006]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61006) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 468.182110] env[61006]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-b8863401-7af7-465b-a63d-ee44f8bcf7e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.198111] env[61006]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61006) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 468.198333] env[61006]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b02fe148-27a3-4b1d-aff0-b0a9c18e7662 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.224013] env[61006]: INFO oslo_vmware.api [-] Successfully established new session; session ID is eb44e. [ 468.224190] env[61006]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.137s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.224672] env[61006]: INFO nova.virt.vmwareapi.driver [None req-826a0a57-3f60-42da-b715-c3851c4c71bb None None] VMware vCenter version: 7.0.3 [ 468.228165] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9c463f-9820-4e70-a5c6-017907d1bdba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.245217] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8186e2fd-58e2-43d8-ab96-fb0488e87099 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.250919] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b27fa08-1fba-4b32-891d-92f7a149f544 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.257334] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab5e3ba-2ba3-47b5-8c60-01b350db7d3b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.270162] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82df8fa3-4ab2-4f8c-8ab7-9f5e1ab76e6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.275948] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc63964-1d51-480f-81f8-095cb136ed4c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.305138] env[61006]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-4ba6725c-1abb-41b2-a0ee-906139433e2f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.310543] env[61006]: DEBUG nova.virt.vmwareapi.driver [None req-826a0a57-3f60-42da-b715-c3851c4c71bb None None] Extension org.openstack.compute already exists. {{(pid=61006) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 468.313230] env[61006]: INFO nova.compute.provider_config [None req-826a0a57-3f60-42da-b715-c3851c4c71bb None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 468.816313] env[61006]: DEBUG nova.context [None req-826a0a57-3f60-42da-b715-c3851c4c71bb None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),44be899a-12ed-4bda-82e0-49b5bed38b08(cell1) {{(pid=61006) load_cells /opt/stack/nova/nova/context.py:464}} [ 468.818534] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 468.818748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 468.819489] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.819911] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Acquiring lock "44be899a-12ed-4bda-82e0-49b5bed38b08" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 468.820108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Lock "44be899a-12ed-4bda-82e0-49b5bed38b08" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 468.821105] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Lock "44be899a-12ed-4bda-82e0-49b5bed38b08" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.840597] env[61006]: INFO dbcounter [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Registered counter for database nova_cell0 [ 468.848412] env[61006]: INFO dbcounter [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Registered counter for database nova_cell1 [ 468.851831] env[61006]: DEBUG oslo_db.sqlalchemy.engines [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61006) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 468.852441] env[61006]: DEBUG oslo_db.sqlalchemy.engines [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61006) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 468.856985] env[61006]: ERROR nova.db.main.api [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 468.856985] env[61006]: result = function(*args, **kwargs) [ 468.856985] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 468.856985] env[61006]: return func(*args, **kwargs) [ 468.856985] env[61006]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 468.856985] env[61006]: result = fn(*args, **kwargs) [ 468.856985] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 468.856985] env[61006]: return f(*args, **kwargs) [ 468.856985] env[61006]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 468.856985] env[61006]: return db.service_get_minimum_version(context, binaries) [ 468.856985] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 468.856985] env[61006]: _check_db_access() [ 468.856985] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 468.856985] env[61006]: stacktrace = ''.join(traceback.format_stack()) [ 468.856985] env[61006]: [ 468.858059] env[61006]: ERROR nova.db.main.api [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 468.858059] env[61006]: result = function(*args, **kwargs) [ 468.858059] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 468.858059] env[61006]: return func(*args, **kwargs) [ 468.858059] env[61006]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 468.858059] env[61006]: result = fn(*args, **kwargs) [ 468.858059] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 468.858059] env[61006]: return f(*args, **kwargs) [ 468.858059] env[61006]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 468.858059] env[61006]: return db.service_get_minimum_version(context, binaries) [ 468.858059] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 468.858059] env[61006]: _check_db_access() [ 468.858059] env[61006]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 468.858059] env[61006]: stacktrace = ''.join(traceback.format_stack()) [ 468.858059] env[61006]: [ 468.858780] env[61006]: WARNING nova.objects.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 468.858780] env[61006]: WARNING nova.objects.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Failed to get minimum service version for cell 44be899a-12ed-4bda-82e0-49b5bed38b08 [ 468.858780] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Acquiring lock "singleton_lock" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 468.858905] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Acquired lock "singleton_lock" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 468.859110] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Releasing lock "singleton_lock" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 468.859442] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Full set of CONF: {{(pid=61006) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 468.859580] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ******************************************************************************** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 468.859705] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Configuration options gathered from: {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 468.859855] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 468.860059] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 468.860187] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ================================================================================ {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 468.860393] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] allow_resize_to_same_host = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.860560] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] arq_binding_timeout = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.860690] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] backdoor_port = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.860813] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] backdoor_socket = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.860971] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] block_device_allocate_retries = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.861149] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] block_device_allocate_retries_interval = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.861315] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cert = self.pem {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.861476] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.861640] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute_monitors = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.861808] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] config_dir = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862007] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] config_drive_format = iso9660 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862160] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862326] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] config_source = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862490] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] console_host = devstack {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862654] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] control_exchange = nova {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862809] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cpu_allocation_ratio = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.862964] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] daemon = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863143] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] debug = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863299] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_access_ip_network_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863457] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_availability_zone = nova {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863610] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_ephemeral_format = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863765] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_green_pool_size = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.863993] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864168] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] default_schedule_zone = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864320] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] disk_allocation_ratio = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864477] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] enable_new_services = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864651] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] enabled_apis = ['osapi_compute'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864808] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] enabled_ssl_apis = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.864964] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] flat_injected = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.865159] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] force_config_drive = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.865342] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] force_raw_images = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.865517] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] graceful_shutdown_timeout = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.865679] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] heal_instance_info_cache_interval = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.865890] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] host = cpu-1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866081] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866246] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866403] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866608] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866769] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_build_timeout = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.866924] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_delete_interval = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867098] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_format = [instance: %(uuid)s] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867271] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_name_template = instance-%08x {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867476] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_usage_audit = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867661] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_usage_audit_period = month {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867831] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.867997] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868176] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] internal_service_availability_zone = internal {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868329] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] key = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868484] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] live_migration_retry_count = 30 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868650] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_color = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868808] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_config_append = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.868968] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869136] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_dir = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869290] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869413] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_options = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869569] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_rotate_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869759] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_rotate_interval_type = days {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.869895] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] log_rotation_type = none {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870042] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870174] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870342] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870503] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870633] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870794] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] long_rpc_timeout = 1800 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.870947] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_concurrent_builds = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871116] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_concurrent_live_migrations = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871271] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_concurrent_snapshots = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871425] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_local_block_devices = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871581] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_logfile_count = 30 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871735] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] max_logfile_size_mb = 200 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.871890] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] maximum_instance_delete_attempts = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872065] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metadata_listen = 0.0.0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872230] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metadata_listen_port = 8775 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872393] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metadata_workers = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872548] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] migrate_max_retries = -1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872710] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] mkisofs_cmd = genisoimage {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.872909] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873094] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] my_ip = 10.180.1.21 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873206] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] network_allocate_retries = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873379] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873542] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873702] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] osapi_compute_listen_port = 8774 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.873926] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] osapi_compute_unique_server_name_scope = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874039] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] osapi_compute_workers = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874204] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] password_length = 12 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874359] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] periodic_enable = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874513] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] periodic_fuzzy_delay = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874676] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] pointer_model = usbtablet {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874836] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] preallocate_images = none {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.874991] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] publish_errors = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875128] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] pybasedir = /opt/stack/nova {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875280] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ram_allocation_ratio = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875468] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rate_limit_burst = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875641] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rate_limit_except_level = CRITICAL {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875797] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rate_limit_interval = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.875951] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reboot_timeout = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.876120] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reclaim_instance_interval = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.876274] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] record = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.876438] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reimage_timeout_per_gb = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.876604] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] report_interval = 120 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.876758] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rescue_timeout = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877029] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reserved_host_cpus = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877106] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reserved_host_disk_mb = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877229] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reserved_host_memory_mb = 512 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877381] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] reserved_huge_pages = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877570] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] resize_confirm_window = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877733] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] resize_fs_using_block_device = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.877888] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] resume_guests_state_on_host_boot = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878064] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878226] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] rpc_response_timeout = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878382] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] run_external_periodic_tasks = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878546] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] running_deleted_instance_action = reap {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878702] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.878856] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] running_deleted_instance_timeout = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879016] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler_instance_sync_interval = 120 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879184] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_down_time = 720 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879350] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] servicegroup_driver = db {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879533] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] shell_completion = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879703] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] shelved_offload_time = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.879882] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] shelved_poll_interval = 3600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880065] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] shutdown_timeout = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880226] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] source_is_ipv6 = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880381] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ssl_only = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880618] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880786] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] sync_power_state_interval = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.880943] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] sync_power_state_pool_size = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881120] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] syslog_log_facility = LOG_USER {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881274] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] tempdir = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881427] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] timeout_nbd = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881591] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] transport_url = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881752] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] update_resources_interval = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.881901] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_cow_images = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882064] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_eventlog = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882218] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_journal = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882369] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_json = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882519] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_rootwrap_daemon = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882672] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_stderr = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882825] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] use_syslog = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.882974] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vcpu_pin_set = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883147] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plugging_is_fatal = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883309] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plugging_timeout = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883470] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] virt_mkfs = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883628] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] volume_usage_poll_interval = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883784] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] watch_log_file = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.883948] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] web = /usr/share/spice-html5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 468.884140] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.884305] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.884466] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.884633] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_concurrency.disable_process_locking = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.885172] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.885388] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.885580] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.885761] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.885934] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.886131] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.886362] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.auth_strategy = keystone {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.compute_link_prefix = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.dhcp_domain = novalocal {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887250] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.enable_instance_password = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887354] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.glance_link_prefix = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887567] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887751] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.887917] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.instance_list_per_project_cells = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888125] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.list_records_by_skipping_down_cells = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888255] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.local_metadata_per_cell = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888448] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.max_limit = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888610] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.metadata_cache_expiration = 15 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888785] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.neutron_default_tenant_id = default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.888956] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.response_validation = warn {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.889147] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.use_neutron_default_nets = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.889312] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.889477] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.889665] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.889912] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.890133] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_dynamic_targets = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.890346] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_jsonfile_path = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.890558] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.890762] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.backend = dogpile.cache.memcached {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.890933] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.backend_argument = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891123] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.config_prefix = cache.oslo {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891297] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.dead_timeout = 60.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891461] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.debug_cache_backend = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891627] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.enable_retry_client = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891790] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.enable_socket_keepalive = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.891958] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.enabled = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892138] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.enforce_fips_mode = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892303] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.expiration_time = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892467] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.hashclient_retry_attempts = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892634] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892795] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_dead_retry = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.892951] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_password = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893128] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893291] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893453] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_pool_maxsize = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893615] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893778] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_sasl_enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.893955] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894136] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894295] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.memcache_username = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894460] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.proxies = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894626] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_db = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894785] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_password = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.894953] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.895140] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.895378] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_server = localhost:6379 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.895638] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_socket_timeout = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.895868] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.redis_username = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.896115] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.retry_attempts = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.896360] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.retry_delay = 0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.896584] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.socket_keepalive_count = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.896812] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.socket_keepalive_idle = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.897057] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.socket_keepalive_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.897306] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.tls_allowed_ciphers = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.897571] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.tls_cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.897826] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.tls_certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898026] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.tls_enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898198] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cache.tls_keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898374] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898559] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.auth_type = password {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898717] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.898891] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899065] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899233] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899397] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.cross_az_attach = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899559] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.debug = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899719] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.endpoint_template = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.899881] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.http_retries = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900055] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900217] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900386] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.os_region_name = RegionOne {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900569] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900842] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cinder.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.900894] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901038] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.cpu_dedicated_set = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901202] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.cpu_shared_set = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901368] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.image_type_exclude_list = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901532] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901696] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.901854] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902026] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902192] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902353] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.resource_provider_association_refresh = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902512] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902673] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.shutdown_retry_interval = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.902849] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903036] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] conductor.workers = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903220] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] console.allowed_origins = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903380] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] console.ssl_ciphers = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903550] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] console.ssl_minimum_version = default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903720] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] consoleauth.enforce_session_timeout = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.903887] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] consoleauth.token_ttl = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904070] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904232] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904394] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904550] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904709] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.904864] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905032] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905194] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905368] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905538] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905699] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.905857] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906021] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906191] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.service_type = accelerator {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906349] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906505] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906661] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906815] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.906993] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.907165] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] cyborg.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.907345] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.backend = sqlalchemy {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.907542] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.connection = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.907716] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.connection_debug = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.907887] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.connection_parameters = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908064] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.connection_recycle_time = 3600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908230] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.connection_trace = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908393] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.db_inc_retry_interval = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908584] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.db_max_retries = 20 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908750] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.db_max_retry_interval = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.908912] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.db_retry_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909087] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.max_overflow = 50 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909252] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.max_pool_size = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909414] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.max_retries = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909584] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909744] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.mysql_wsrep_sync_wait = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.909901] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.pool_timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910074] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.retry_interval = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910235] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.slave_connection = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910395] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.sqlite_synchronous = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910555] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] database.use_db_reconnect = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910734] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.backend = sqlalchemy {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.910900] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.connection = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911078] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.connection_debug = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911252] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.connection_parameters = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911415] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.connection_recycle_time = 3600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911579] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.connection_trace = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911742] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.db_inc_retry_interval = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.911892] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.db_max_retries = 20 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912062] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.db_max_retry_interval = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912227] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.db_retry_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912385] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.max_overflow = 50 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912544] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.max_pool_size = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912706] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.max_retries = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.912871] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913201] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.pool_timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913360] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.retry_interval = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913516] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.slave_connection = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913675] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] api_database.sqlite_synchronous = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.913851] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] devices.enabled_mdev_types = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914042] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914218] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914381] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ephemeral_storage_encryption.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914543] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914715] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.api_servers = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.914876] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915047] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915212] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915391] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915561] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915727] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.debug = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.915891] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.default_trusted_certificate_ids = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916066] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.enable_certificate_validation = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916232] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.enable_rbd_download = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916394] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916551] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916724] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.916881] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917045] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917208] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.num_retries = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917375] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.rbd_ceph_conf = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917572] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.rbd_connect_timeout = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917744] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.rbd_pool = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.917912] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.rbd_user = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918089] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918251] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918496] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918589] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.service_type = image {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918739] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.918904] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919071] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919230] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919407] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919577] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.verify_glance_signatures = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919759] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] glance.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.919926] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] guestfs.debug = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.920107] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] mks.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.920469] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.920661] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.manager_interval = 2400 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.920831] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.precache_concurrency = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921006] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.remove_unused_base_images = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921187] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921355] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921531] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] image_cache.subdirectory_name = _base {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921707] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.api_max_retries = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.921873] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.api_retry_interval = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922207] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.auth_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922368] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922528] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922714] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.922885] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.conductor_group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923056] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923219] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923375] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923536] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923695] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.923852] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924013] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924195] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.peer_list = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924350] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924507] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924671] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.serial_console_state_timeout = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924830] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.924998] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.service_type = baremetal {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.925172] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.shard = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.925345] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.925549] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.925727] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.925888] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926089] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926255] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ironic.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926435] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926610] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] key_manager.fixed_key = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926875] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.926951] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.barbican_api_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927127] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.barbican_endpoint = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927300] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.barbican_endpoint_type = public {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927478] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.barbican_region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927649] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927810] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.927972] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.928148] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.928307] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.928479] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.number_of_retries = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.928707] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.retry_delay = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.928883] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.send_service_user_token = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929061] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929224] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929386] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.verify_ssl = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929545] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican.verify_ssl_path = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929712] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.929874] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.auth_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930045] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930208] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930371] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930530] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930689] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.930850] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931015] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] barbican_service_user.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931190] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.approle_role_id = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931347] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.approle_secret_id = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931532] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.kv_mountpoint = secret {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931710] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.kv_path = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.931882] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.kv_version = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932047] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.namespace = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932210] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.root_token_id = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932368] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.ssl_ca_crt_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932536] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.timeout = 60.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932700] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.use_ssl = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.932868] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933049] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933218] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.auth_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933377] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933537] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933701] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.933858] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934024] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934187] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934345] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934500] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934684] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934852] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.934998] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.935171] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.935337] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.935528] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.service_type = identity {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.935693] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.935851] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936013] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936178] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936357] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936513] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] keystone.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936713] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.connection_uri = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.936875] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_mode = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937051] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937223] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_models = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937396] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_power_governor_high = performance {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937585] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937751] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_power_management = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.937924] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938106] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.device_detach_attempts = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938272] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.device_detach_timeout = 20 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938442] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.disk_cachemodes = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938601] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.disk_prefix = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938762] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.enabled_perf_events = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.938921] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.file_backed_memory = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939097] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.gid_maps = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939256] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.hw_disk_discard = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939413] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.hw_machine_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939579] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_rbd_ceph_conf = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939747] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.939910] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940088] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_rbd_glance_store_name = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940258] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_rbd_pool = rbd {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940429] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_type = default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940588] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.images_volume_group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940748] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.inject_key = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.940908] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.inject_partition = -2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941079] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.inject_password = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941241] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.iscsi_iface = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941400] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.iser_use_multipath = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941565] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941723] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.941882] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_downtime = 500 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942071] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942216] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942375] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_inbound_addr = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942535] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942701] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.942859] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_scheme = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943046] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_timeout_action = abort {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943212] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_tunnelled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943371] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_uri = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943541] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.live_migration_with_native_tls = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943704] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.max_queues = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.943864] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.944119] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.944288] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.nfs_mount_options = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.944596] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.944770] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.944934] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945109] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945274] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945456] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_pcie_ports = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945628] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945796] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.pmem_namespaces = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.945950] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.quobyte_client_cfg = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.946252] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.946425] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.946591] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.946765] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.946909] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rbd_secret_uuid = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947079] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rbd_user = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947245] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947412] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947602] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rescue_image_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947765] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rescue_kernel_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.947921] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rescue_ramdisk_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.948101] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.948283] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.rx_queue_size = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.948467] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.smbfs_mount_options = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.948743] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.948913] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.snapshot_compression = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949085] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.snapshot_image_format = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949306] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949470] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.sparse_logical_volumes = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949634] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.swtpm_enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949800] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.swtpm_group = tss {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.949966] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.swtpm_user = tss {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950153] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.sysinfo_serial = unique {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950309] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.tb_cache_size = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950463] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.tx_queue_size = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950626] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.uid_maps = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950785] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.use_virtio_for_bridges = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.950954] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.virt_type = kvm {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951137] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.volume_clear = zero {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951299] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.volume_clear_size = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951464] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.volume_use_multipath = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951623] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_cache_path = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951790] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.951956] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.952170] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.952302] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.952572] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.952748] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.vzstorage_mount_user = stack {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.952911] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953098] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953275] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.auth_type = password {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953437] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953595] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953756] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.953911] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954080] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954255] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.default_floating_pool = public {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954412] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954570] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.extension_sync_interval = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954731] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.http_retries = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.954889] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955059] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955221] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955407] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955578] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955750] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.ovs_bridge = br-int {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.955917] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.physnets = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956098] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.region_name = RegionOne {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956260] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956429] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.service_metadata_proxy = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956591] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956758] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.service_type = network {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.956921] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957091] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957252] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957408] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957620] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957786] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] neutron.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.957958] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] notifications.bdms_in_notifications = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.958151] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] notifications.default_level = INFO {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.958329] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] notifications.notification_format = unversioned {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.958516] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] notifications.notify_on_state_change = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.958706] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.958885] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] pci.alias = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959070] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] pci.device_spec = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959239] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] pci.report_in_placement = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959408] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959581] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.auth_type = password {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959749] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.959907] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960074] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960240] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960395] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960551] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960707] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.default_domain_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.960860] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.default_domain_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961025] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.domain_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961187] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.domain_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961344] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961504] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961660] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961812] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.961964] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962145] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.password = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962303] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.project_domain_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962467] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.project_domain_name = Default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962636] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.project_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962810] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.project_name = service {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.962978] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.region_name = RegionOne {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963155] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963315] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963482] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.service_type = placement {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963645] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963801] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.963988] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964166] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.system_scope = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964321] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964485] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.trust_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964647] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.user_domain_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964812] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.user_domain_name = Default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.964968] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.user_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.965155] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.username = nova {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.965345] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.965523] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] placement.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.965707] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.cores = 20 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.965871] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.count_usage_from_placement = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966055] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966234] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.injected_file_content_bytes = 10240 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966400] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.injected_file_path_length = 255 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966566] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.injected_files = 5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966736] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.instances = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.966907] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.key_pairs = 100 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967076] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.metadata_items = 128 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967244] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.ram = 51200 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967402] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.recheck_quota = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967592] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.server_group_members = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967762] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] quota.server_groups = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.967930] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968103] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968265] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.image_metadata_prefilter = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968427] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968589] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.max_attempts = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968749] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.max_placement_results = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.968908] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969074] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969236] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969403] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] scheduler.workers = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969571] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969742] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.969918] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970095] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970263] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970425] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970587] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970779] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.970948] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.host_subset_size = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971125] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971284] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971446] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971611] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.isolated_hosts = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971771] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.isolated_images = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.971934] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972105] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972269] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972431] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.pci_in_placement = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972592] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972750] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.972909] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973077] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973240] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973399] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973558] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.track_instance_changes = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973734] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.973904] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metrics.required = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.974079] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metrics.weight_multiplier = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.974242] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.974403] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] metrics.weight_setting = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.974719] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.974895] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975083] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.port_range = 10000:20000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975259] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975447] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975623] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] serial_console.serialproxy_port = 6083 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975798] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.975973] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.auth_type = password {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976147] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976308] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976467] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976629] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976786] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.976953] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.send_service_user_token = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.977128] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.977299] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] service_user.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.977487] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.agent_enabled = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.977662] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.977982] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.978189] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.978359] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.html5proxy_port = 6082 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.978544] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.image_compression = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.978711] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.jpeg_compression = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.978866] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.playback_compression = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979037] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.require_secure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979209] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.server_listen = 127.0.0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979376] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979533] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.streaming_mode = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979691] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] spice.zlib_compression = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.979858] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] upgrade_levels.baseapi = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980037] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] upgrade_levels.compute = auto {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980195] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] upgrade_levels.conductor = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980353] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] upgrade_levels.scheduler = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980533] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980717] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.980878] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981048] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981213] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981375] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981533] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981696] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.981853] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vendordata_dynamic_auth.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982033] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.api_retry_count = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982198] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.ca_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982369] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982537] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.cluster_name = testcl1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982702] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.connection_pool_size = 10 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.982858] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.console_delay_seconds = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983036] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.datastore_regex = ^datastore.* {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983253] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983416] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.host_password = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983581] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.host_port = 443 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983749] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.host_username = administrator@vsphere.local {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.983916] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.insecure = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984088] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.integration_bridge = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984256] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.maximum_objects = 100 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984414] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.pbm_default_policy = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984575] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.pbm_enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984733] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.pbm_wsdl_location = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.984899] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985068] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.serial_port_proxy_uri = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985228] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.serial_port_service_uri = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985411] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.task_poll_interval = 0.5 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985594] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.use_linked_clone = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985764] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.vnc_keymap = en-us {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.985926] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.vnc_port = 5900 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.986101] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vmware.vnc_port_total = 10000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.986290] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.auth_schemes = ['none'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.986464] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.986766] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.986950] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.987161] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.novncproxy_port = 6080 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.987328] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.server_listen = 127.0.0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.987527] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.987702] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.vencrypt_ca_certs = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.987862] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.vencrypt_client_cert = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988031] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vnc.vencrypt_client_key = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988214] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988379] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_deep_image_inspection = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988552] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988728] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.988889] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989060] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.disable_rootwrap = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989227] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.enable_numa_live_migration = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989387] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989548] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989723] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.989886] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.libvirt_disable_apic = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990056] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990224] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990384] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990543] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990705] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.990885] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991060] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991225] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991384] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991545] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991731] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.991899] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.client_socket_timeout = 900 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992077] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.default_pool_size = 1000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992247] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.keep_alive = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992412] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.max_header_line = 16384 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992572] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992731] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.ssl_ca_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.992889] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.ssl_cert_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.993057] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.ssl_key_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.993223] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.tcp_keepidle = 600 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.993398] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.993562] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] zvm.ca_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.993720] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] zvm.cloud_connector_url = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.994015] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.994195] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] zvm.reachable_timeout = 300 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.994375] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.enforce_new_defaults = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.994759] env[61006]: WARNING oslo_config.cfg [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 468.994943] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.enforce_scope = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.995132] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.policy_default_rule = default {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.995320] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.995505] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.policy_file = policy.yaml {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.995679] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.995853] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996021] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996181] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996341] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996508] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996703] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.996893] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.connection_string = messaging:// {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997074] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.enabled = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997265] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.es_doc_type = notification {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997410] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.es_scroll_size = 10000 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997607] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.es_scroll_time = 2m {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997776] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.filter_error_trace = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.997946] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.hmac_keys = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998128] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.sentinel_service_name = mymaster {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998293] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.socket_timeout = 0.1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998456] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.trace_requests = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998615] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler.trace_sqlalchemy = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998791] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler_jaeger.process_tags = {} {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.998951] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler_jaeger.service_name_prefix = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999127] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] profiler_otlp.service_name_prefix = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999296] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] remote_debug.host = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999456] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] remote_debug.port = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999649] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999830] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 468.999992] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000167] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000332] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000492] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000654] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000814] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.000974] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001157] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001317] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001493] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001662] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001830] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.001998] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.002181] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.002342] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.002515] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.002696] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.002872] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003053] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003222] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003385] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003551] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003713] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.003872] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004041] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004205] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004371] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004535] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004706] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.004877] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005049] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005222] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005391] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005552] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005758] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.005934] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_notifications.retry = -1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006137] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006315] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006485] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.auth_section = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006647] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.auth_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006804] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.cafile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.006961] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.certfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007138] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.collect_timing = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007296] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.connect_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007476] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.connect_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007641] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.endpoint_id = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007800] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.endpoint_override = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.007960] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.insecure = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008132] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.keyfile = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008290] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.max_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008460] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.min_version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008656] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.region_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008826] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.retriable_status_codes = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.008988] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.service_name = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009160] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.service_type = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009322] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.split_loggers = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009480] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.status_code_retries = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009638] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.status_code_retry_delay = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009815] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.timeout = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.009984] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.valid_interfaces = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010157] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_limit.version = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010324] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_reports.file_event_handler = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010485] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010644] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] oslo_reports.log_dir = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010813] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.010969] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011138] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011301] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011470] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011652] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011824] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.011984] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012156] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012322] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012481] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012639] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] vif_plug_ovs_privileged.user = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012809] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.012987] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.013177] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.013348] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.013518] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.013685] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.013851] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014017] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014200] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014372] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.isolate_vif = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014557] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014792] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.014990] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.015189] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.015355] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] os_vif_ovs.per_port_bridge = False {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.015526] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.capabilities = [21] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.015688] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.015846] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.helper_command = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016018] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016189] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016347] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] privsep_osbrick.user = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016520] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016682] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.group = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.016841] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.helper_command = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.017013] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.017189] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.017346] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] nova_sys_admin.user = None {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.017498] env[61006]: DEBUG oslo_service.service [None req-4fe74094-ebfb-4afe-b60a-0001d989e2f8 None None] ******************************************************************************** {{(pid=61006) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 469.018023] env[61006]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 469.521316] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Getting list of instances from cluster (obj){ [ 469.521316] env[61006]: value = "domain-c8" [ 469.521316] env[61006]: _type = "ClusterComputeResource" [ 469.521316] env[61006]: } {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 469.522555] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec813434-e56f-4402-a58d-bb0001854d86 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.531565] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Got total of 0 instances {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 469.532115] env[61006]: WARNING nova.virt.vmwareapi.driver [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 469.532593] env[61006]: INFO nova.virt.node [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Generated node identity 3360045e-46ab-4f2d-9377-dd481ab3cd53 [ 469.532841] env[61006]: INFO nova.virt.node [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Wrote node identity 3360045e-46ab-4f2d-9377-dd481ab3cd53 to /opt/stack/data/n-cpu-1/compute_id [ 470.035496] env[61006]: WARNING nova.compute.manager [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Compute nodes ['3360045e-46ab-4f2d-9377-dd481ab3cd53'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 471.040913] env[61006]: INFO nova.compute.manager [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 472.046491] env[61006]: WARNING nova.compute.manager [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 472.046847] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.046947] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.047077] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.047233] env[61006]: DEBUG nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 472.048249] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc1f576-2dc2-4907-994e-090e447beb64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.056466] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a0df6e-93b0-407a-91b0-3747ac48b587 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.069784] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce85fc63-a9b0-4c42-be22-4055ef2c799d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.076148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a251511-9ac8-46c5-8f1d-07b640ac78b4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.104897] env[61006]: DEBUG nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181458MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 472.105053] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.105242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.608225] env[61006]: WARNING nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] No compute node record for cpu-1:3360045e-46ab-4f2d-9377-dd481ab3cd53: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 3360045e-46ab-4f2d-9377-dd481ab3cd53 could not be found. [ 473.112160] env[61006]: INFO nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 3360045e-46ab-4f2d-9377-dd481ab3cd53 [ 474.620248] env[61006]: DEBUG nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 474.620471] env[61006]: DEBUG nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 474.770821] env[61006]: INFO nova.scheduler.client.report [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] [req-83355190-f348-4ae1-9f61-bc687ca1daac] Created resource provider record via placement API for resource provider with UUID 3360045e-46ab-4f2d-9377-dd481ab3cd53 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 474.787296] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4abd5cb-dd01-41f7-ac6e-790817ef36b4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.795453] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8516a0c8-3264-4cd5-adfb-716a1795954a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.825053] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1c2c9e-b54f-4f3e-9e6e-d656840388b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.831668] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44268ee6-69bf-4dd4-b0de-bb80329c216e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.843908] env[61006]: DEBUG nova.compute.provider_tree [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 475.381936] env[61006]: DEBUG nova.scheduler.client.report [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 475.382193] env[61006]: DEBUG nova.compute.provider_tree [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 0 to 1 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 475.382335] env[61006]: DEBUG nova.compute.provider_tree [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 475.433720] env[61006]: DEBUG nova.compute.provider_tree [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 1 to 2 during operation: update_traits {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 475.938260] env[61006]: DEBUG nova.compute.resource_tracker [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 475.938640] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.833s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.938640] env[61006]: DEBUG nova.service [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Creating RPC server for service compute {{(pid=61006) start /opt/stack/nova/nova/service.py:186}} [ 475.956051] env[61006]: DEBUG nova.service [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] Join ServiceGroup membership for this service compute {{(pid=61006) start /opt/stack/nova/nova/service.py:203}} [ 475.956232] env[61006]: DEBUG nova.servicegroup.drivers.db [None req-d76c286d-c5f7-4f35-8036-0a0d1b7a5fed None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61006) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 510.233502] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquiring lock "f3feb305-2737-4398-84d3-311a4f9c42fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.233849] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "f3feb305-2737-4398-84d3-311a4f9c42fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.736540] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.285328] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.285692] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.287405] env[61006]: INFO nova.compute.claims [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 512.176374] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.176561] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.374849] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4905abff-e13b-4bbc-ba22-ecf2f810b329 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.384357] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a625c470-7f28-45ea-97dd-dcec79632d85 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.422241] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0a7188-75c6-45c5-bfd8-b41d9366eb89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.430557] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb6eaf3-8878-4676-a3c3-07ff44dc681e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.444952] env[61006]: DEBUG nova.compute.provider_tree [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.685629] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 512.948335] env[61006]: DEBUG nova.scheduler.client.report [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 513.228312] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.455434] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.455772] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.458567] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.230s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.466127] env[61006]: INFO nova.compute.claims [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.849627] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquiring lock "3769d396-ef6a-436b-ab75-faf66e4da813" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.850024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "3769d396-ef6a-436b-ab75-faf66e4da813" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.972144] env[61006]: DEBUG nova.compute.utils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.973702] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 513.974197] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 514.354214] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.481894] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.529212] env[61006]: DEBUG nova.policy [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69069098ed89483e9353f1977680ad46', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d45789a9924471aabcbccd9995b32c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 514.576028] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79f36fe-0f4a-42db-86ff-aea80c77576c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.586580] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cb5098-3317-4bed-a99b-f0473b3a501f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.625136] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a539b720-bc0f-4417-b4f8-2ca63d8e2313 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.634902] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74591d7-8bdf-40b7-935b-3017449b467d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.650440] env[61006]: DEBUG nova.compute.provider_tree [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.842673] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.843179] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.891539] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.155161] env[61006]: DEBUG nova.scheduler.client.report [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.348473] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.413296] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.413455] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.494635] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.518020] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Successfully created port: c55b3184-7509-4a17-bd6d-de59ec38bcba {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 515.540098] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.540098] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.540098] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.540232] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.540232] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.540232] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.540357] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.540532] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.540909] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.541115] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.541361] env[61006]: DEBUG nova.virt.hardware [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.542211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502bdf6e-0aa0-4502-a921-409bfe33c9f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.553080] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e518940-dc05-4d97-9cc4-2254e379c662 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.577553] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe4e715-46b7-4afb-8a11-b44265f6d9ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.660589] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.661860] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 515.670776] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.779s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.676281] env[61006]: INFO nova.compute.claims [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.877933] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.917177] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.942859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.943410] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.071045] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "0ace1533-286c-42e2-afde-e99c77fdd747" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.071559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "0ace1533-286c-42e2-afde-e99c77fdd747" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.176730] env[61006]: DEBUG nova.compute.utils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.178312] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 516.178966] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 516.365804] env[61006]: DEBUG nova.policy [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 516.446858] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.455927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.574890] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.683705] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 516.849750] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1103d71-0aca-40c9-867b-c57efd1b1a91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.860561] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454bf5ef-b193-452a-b9ed-bd17beb07075 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.905724] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ba57ee-f260-4d4a-b765-48e074c6c3ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.916020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979bf9e5-0b98-43e5-b5a2-4e9aadbf4410 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.928832] env[61006]: DEBUG nova.compute.provider_tree [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.974673] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.100582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.101415] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Successfully created port: 03ae18ba-2881-4fe6-8eb1-5c6066c15be4 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 517.432830] env[61006]: DEBUG nova.scheduler.client.report [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.696302] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 517.734209] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.734209] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.734209] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.734427] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.734427] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.734481] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.734742] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.734903] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.735164] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.735784] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.736077] env[61006]: DEBUG nova.virt.hardware [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.737352] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc39874e-8245-453a-92a4-9bb5702da9cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.746653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705ac525-0b4a-4c19-9e56-9848c66219bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.938031] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.938604] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.941696] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.064s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.943855] env[61006]: INFO nova.compute.claims [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.100275] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquiring lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.100503] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.451226] env[61006]: DEBUG nova.compute.utils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.454774] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 518.456452] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 518.603796] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.687198] env[61006]: DEBUG nova.policy [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e22c8ac236544e2eb6d7251a220e4ac7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70b0c2ee912b4bfebe0ae3e792a49efb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 518.962580] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.126095] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.130495] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789851dd-0b6d-4ea8-98f4-ec07653fb64a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.143705] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1425e1-61ee-449c-945b-504681f6fed1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.179274] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0d3d65-d529-4ffd-8f23-1d9f9e825f28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.187848] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90de5aa-f01d-4ba7-a5e3-3122f970e47f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.205518] env[61006]: DEBUG nova.compute.provider_tree [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.709746] env[61006]: DEBUG nova.scheduler.client.report [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.980011] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.012017] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.012017] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.012017] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.012228] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.012228] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.012418] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.012969] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.015097] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.015097] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.015097] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.015097] env[61006]: DEBUG nova.virt.hardware [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.015097] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98f4b18-ddfe-46cb-87c2-7df71999484d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.026141] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef4b29b-3c65-4e93-b447-6c3a3f01d4d8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.207268] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Successfully created port: a24e2295-0c85-4318-a371-94d32be9aada {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 520.218163] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.218163] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.219176] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.764s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.221878] env[61006]: INFO nova.compute.claims [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.731051] env[61006]: DEBUG nova.compute.utils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.738578] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 521.243457] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.286835] env[61006]: ERROR nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 521.286835] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.286835] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.286835] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.286835] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.286835] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.286835] env[61006]: ERROR nova.compute.manager raise self.value [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.286835] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 521.286835] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.286835] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 521.287425] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.287425] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 521.287425] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 521.287425] env[61006]: ERROR nova.compute.manager [ 521.287923] env[61006]: Traceback (most recent call last): [ 521.288536] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 521.288536] env[61006]: listener.cb(fileno) [ 521.288725] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.288725] env[61006]: result = function(*args, **kwargs) [ 521.288821] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.288821] env[61006]: return func(*args, **kwargs) [ 521.288890] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.288890] env[61006]: raise e [ 521.289058] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.289058] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 521.289120] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.289120] env[61006]: created_port_ids = self._update_ports_for_instance( [ 521.289188] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.289188] env[61006]: with excutils.save_and_reraise_exception(): [ 521.289254] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.289254] env[61006]: self.force_reraise() [ 521.289331] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.289331] env[61006]: raise self.value [ 521.289398] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.289398] env[61006]: updated_port = self._update_port( [ 521.289464] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.289464] env[61006]: _ensure_no_port_binding_failure(port) [ 521.289536] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.289536] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 521.289604] env[61006]: nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 521.289652] env[61006]: Removing descriptor: 15 [ 521.292157] env[61006]: ERROR nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Traceback (most recent call last): [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] yield resources [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.driver.spawn(context, instance, image_meta, [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] vm_ref = self.build_virtual_machine(instance, [ 521.292157] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] for vif in network_info: [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self._sync_wrapper(fn, *args, **kwargs) [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.wait() [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self[:] = self._gt.wait() [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self._exit_event.wait() [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 521.292530] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] result = hub.switch() [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self.greenlet.switch() [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] result = function(*args, **kwargs) [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return func(*args, **kwargs) [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise e [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] nwinfo = self.network_api.allocate_for_instance( [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] created_port_ids = self._update_ports_for_instance( [ 521.292889] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] with excutils.save_and_reraise_exception(): [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.force_reraise() [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise self.value [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] updated_port = self._update_port( [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] _ensure_no_port_binding_failure(port) [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise exception.PortBindingFailed(port_id=port['id']) [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 521.293287] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] [ 521.295668] env[61006]: INFO nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Terminating instance [ 521.302058] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquiring lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.302058] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquired lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.302058] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 521.489885] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0db1d04-d7d2-4d1c-b6a8-b6ccba2f37b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.502235] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1406a7-4b2a-4b6f-b2be-ba31fd0395d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.538564] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ca52ca-e2b7-4c00-957b-61a24aa0a92a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.547043] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ded5dc-27c9-4987-8ac3-cec59a3a0a7f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.564645] env[61006]: DEBUG nova.compute.provider_tree [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.679953] env[61006]: ERROR nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 521.679953] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.679953] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.679953] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.679953] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.679953] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.679953] env[61006]: ERROR nova.compute.manager raise self.value [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.679953] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 521.679953] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.679953] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 521.680402] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.680402] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 521.680402] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 521.680402] env[61006]: ERROR nova.compute.manager [ 521.680402] env[61006]: Traceback (most recent call last): [ 521.680402] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 521.680402] env[61006]: listener.cb(fileno) [ 521.680402] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.680402] env[61006]: result = function(*args, **kwargs) [ 521.680402] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.680402] env[61006]: return func(*args, **kwargs) [ 521.680402] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.680402] env[61006]: raise e [ 521.680402] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.680402] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 521.680402] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.680402] env[61006]: created_port_ids = self._update_ports_for_instance( [ 521.680402] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.680402] env[61006]: with excutils.save_and_reraise_exception(): [ 521.680402] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.680402] env[61006]: self.force_reraise() [ 521.680402] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.680402] env[61006]: raise self.value [ 521.680402] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.680402] env[61006]: updated_port = self._update_port( [ 521.680402] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.680402] env[61006]: _ensure_no_port_binding_failure(port) [ 521.680402] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.680402] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 521.681208] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 521.681208] env[61006]: Removing descriptor: 16 [ 521.681208] env[61006]: ERROR nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Traceback (most recent call last): [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] yield resources [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.driver.spawn(context, instance, image_meta, [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 521.681208] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] vm_ref = self.build_virtual_machine(instance, [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] for vif in network_info: [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self._sync_wrapper(fn, *args, **kwargs) [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.wait() [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self[:] = self._gt.wait() [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self._exit_event.wait() [ 521.681565] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] result = hub.switch() [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self.greenlet.switch() [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] result = function(*args, **kwargs) [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return func(*args, **kwargs) [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise e [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] nwinfo = self.network_api.allocate_for_instance( [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.682195] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] created_port_ids = self._update_ports_for_instance( [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] with excutils.save_and_reraise_exception(): [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.force_reraise() [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise self.value [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] updated_port = self._update_port( [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] _ensure_no_port_binding_failure(port) [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.682495] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise exception.PortBindingFailed(port_id=port['id']) [ 521.682816] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 521.682816] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] [ 521.682816] env[61006]: INFO nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Terminating instance [ 521.685241] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.686691] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.686691] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 521.867212] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.979962] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.071075] env[61006]: DEBUG nova.scheduler.client.report [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.207983] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.268573] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.308640] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.308985] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.309156] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.309350] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.309496] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.309639] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.309846] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.310024] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.310175] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.310616] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.310616] env[61006]: DEBUG nova.virt.hardware [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.314271] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188f7d76-fdca-44df-8c55-81b7d42cae88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.321116] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49979046-9696-464b-a489-d828738abbda {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.338471] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 522.349550] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.349743] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5cbe28c-6202-4b2d-9912-d43892f76191 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.363647] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created folder: OpenStack in parent group-v4. [ 522.365302] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating folder: Project (901b887f6b234ad081eb2615c1c4fe36). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.365302] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91c339ba-a95d-4b02-bfbf-cec93b4dd1a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.374294] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created folder: Project (901b887f6b234ad081eb2615c1c4fe36) in parent group-v285275. [ 522.374478] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating folder: Instances. Parent ref: group-v285276. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.374820] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1557e975-07d4-418b-95ab-f586d5d0dab1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.387300] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created folder: Instances in parent group-v285276. [ 522.387300] env[61006]: DEBUG oslo.service.loopingcall [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.387300] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 522.387300] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-daeb45b2-02d0-4132-9cf9-f2c9770b36e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.412303] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 522.412303] env[61006]: value = "task-1336810" [ 522.412303] env[61006]: _type = "Task" [ 522.412303] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.421100] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336810, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.485781] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Releasing lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.487376] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 522.487376] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 522.487376] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f513454-9a7d-426b-a3f3-5e39500c474c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.498482] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ceeb120-46a2-4c0e-9ed2-4dbc5da2fd21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.519611] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquiring lock "fa649a9c-8745-4188-80b8-09d62cd882e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.520748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "fa649a9c-8745-4188-80b8-09d62cd882e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.535020] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f3feb305-2737-4398-84d3-311a4f9c42fc could not be found. [ 522.535020] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 522.535020] env[61006]: INFO nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 522.535020] env[61006]: DEBUG oslo.service.loopingcall [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.535020] env[61006]: DEBUG nova.compute.manager [-] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 522.535020] env[61006]: DEBUG nova.network.neutron [-] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 522.562060] env[61006]: DEBUG nova.network.neutron [-] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.582119] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.583837] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.584196] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.588682] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.614s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.591905] env[61006]: INFO nova.compute.claims [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.881940] env[61006]: DEBUG nova.compute.manager [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Received event network-changed-c55b3184-7509-4a17-bd6d-de59ec38bcba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 522.881940] env[61006]: DEBUG nova.compute.manager [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Refreshing instance network info cache due to event network-changed-c55b3184-7509-4a17-bd6d-de59ec38bcba. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 522.881940] env[61006]: DEBUG oslo_concurrency.lockutils [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] Acquiring lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.882169] env[61006]: DEBUG oslo_concurrency.lockutils [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] Acquired lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.882273] env[61006]: DEBUG nova.network.neutron [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Refreshing network info cache for port c55b3184-7509-4a17-bd6d-de59ec38bcba {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 522.922674] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336810, 'name': CreateVM_Task, 'duration_secs': 0.296349} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.922840] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 522.923836] env[61006]: DEBUG oslo_vmware.service [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3994d3c3-c37a-4203-a128-c81f125b8438 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.930131] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.930131] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.930515] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 522.930995] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dcddd3a-2a9f-46ff-b2f4-7c4bf3711df2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.936935] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 522.936935] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521435a5-a578-c939-44e5-746582e44622" [ 522.936935] env[61006]: _type = "Task" [ 522.936935] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.948888] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521435a5-a578-c939-44e5-746582e44622, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.022061] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.065325] env[61006]: DEBUG nova.network.neutron [-] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.091122] env[61006]: DEBUG nova.compute.utils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.092088] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.092495] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 523.093222] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 523.093714] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.093875] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.098677] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-532652aa-5bdf-4c4f-b213-c0655d3c4b18 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.110233] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a428eeb-1f59-4f04-a0d1-4e62c2f783c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.138990] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a4d93990-9591-4bb5-8e26-d8a57807d8d0 could not be found. [ 523.139296] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 523.139515] env[61006]: INFO nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 523.139726] env[61006]: DEBUG oslo.service.loopingcall [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.139874] env[61006]: DEBUG nova.compute.manager [-] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 523.139998] env[61006]: DEBUG nova.network.neutron [-] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.180300] env[61006]: DEBUG nova.network.neutron [-] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.224308] env[61006]: DEBUG nova.policy [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2a55b84c0934089b626f47127f9fbac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0032cec34cea4502862dcf3539be253b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.271187] env[61006]: DEBUG nova.compute.manager [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Received event network-changed-03ae18ba-2881-4fe6-8eb1-5c6066c15be4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 523.271187] env[61006]: DEBUG nova.compute.manager [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Refreshing instance network info cache due to event network-changed-03ae18ba-2881-4fe6-8eb1-5c6066c15be4. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 523.271187] env[61006]: DEBUG oslo_concurrency.lockutils [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] Acquiring lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.271757] env[61006]: DEBUG oslo_concurrency.lockutils [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] Acquired lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.271757] env[61006]: DEBUG nova.network.neutron [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Refreshing network info cache for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 523.446812] env[61006]: DEBUG nova.network.neutron [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.459614] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.459863] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 523.460104] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.460762] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.460762] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 523.460943] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43d2e0d8-2a94-4ca5-abab-31f202e70fb3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.482214] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 523.482214] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 523.482214] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa0467c-5ca9-4c7a-b6ff-7d397446e753 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.493783] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7cf429d-b0a8-450b-9b71-eb4508be76a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.501212] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 523.501212] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52eabfb3-8711-0daa-44dd-ccbb8cb04747" [ 523.501212] env[61006]: _type = "Task" [ 523.501212] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.511297] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52eabfb3-8711-0daa-44dd-ccbb8cb04747, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.549239] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.572129] env[61006]: INFO nova.compute.manager [-] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Took 1.04 seconds to deallocate network for instance. [ 523.576675] env[61006]: DEBUG nova.compute.claims [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 523.576818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.594639] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.641318] env[61006]: ERROR nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 523.641318] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.641318] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.641318] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.641318] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.641318] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.641318] env[61006]: ERROR nova.compute.manager raise self.value [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.641318] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.641318] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.641318] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.641762] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.641762] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.641762] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 523.641762] env[61006]: ERROR nova.compute.manager [ 523.641762] env[61006]: Traceback (most recent call last): [ 523.641762] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.641762] env[61006]: listener.cb(fileno) [ 523.641762] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.641762] env[61006]: result = function(*args, **kwargs) [ 523.641762] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.641762] env[61006]: return func(*args, **kwargs) [ 523.641762] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.641762] env[61006]: raise e [ 523.641762] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.641762] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 523.641762] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.641762] env[61006]: created_port_ids = self._update_ports_for_instance( [ 523.641762] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.641762] env[61006]: with excutils.save_and_reraise_exception(): [ 523.641762] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.641762] env[61006]: self.force_reraise() [ 523.641762] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.641762] env[61006]: raise self.value [ 523.641762] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.641762] env[61006]: updated_port = self._update_port( [ 523.641762] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.641762] env[61006]: _ensure_no_port_binding_failure(port) [ 523.641762] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.641762] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.642548] env[61006]: nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 523.642548] env[61006]: Removing descriptor: 17 [ 523.642548] env[61006]: ERROR nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Traceback (most recent call last): [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] yield resources [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.driver.spawn(context, instance, image_meta, [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.642548] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] vm_ref = self.build_virtual_machine(instance, [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] for vif in network_info: [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self._sync_wrapper(fn, *args, **kwargs) [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.wait() [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self[:] = self._gt.wait() [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self._exit_event.wait() [ 523.643053] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] result = hub.switch() [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self.greenlet.switch() [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] result = function(*args, **kwargs) [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return func(*args, **kwargs) [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise e [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] nwinfo = self.network_api.allocate_for_instance( [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.643459] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] created_port_ids = self._update_ports_for_instance( [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] with excutils.save_and_reraise_exception(): [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.force_reraise() [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise self.value [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] updated_port = self._update_port( [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] _ensure_no_port_binding_failure(port) [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.643994] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise exception.PortBindingFailed(port_id=port['id']) [ 523.644505] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 523.644505] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] [ 523.644505] env[61006]: INFO nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Terminating instance [ 523.644785] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquiring lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.644944] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquired lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.645117] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.655963] env[61006]: DEBUG nova.network.neutron [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.684703] env[61006]: DEBUG nova.network.neutron [-] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.811038] env[61006]: DEBUG nova.network.neutron [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.929813] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93fb762d-c7a1-4574-9298-61a1c8ccb41c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.944069] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0844fa61-f113-4034-873e-af2f67202c53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.975415] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7cd6e6-1374-4e69-9b8a-62f245e86ac2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.983907] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e80a38-8bf1-4209-b605-144d1dcebae4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.006464] env[61006]: DEBUG nova.compute.provider_tree [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.008541] env[61006]: DEBUG nova.network.neutron [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.013837] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Successfully created port: 3518a742-741e-4531-9e63-a8e3e8d13f1e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.026061] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 524.026352] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating directory with path [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.026531] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-171de39b-2d49-45e0-b3bd-b75efec287f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.048193] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created directory with path [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.049309] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Fetch image to [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 524.049450] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Downloading image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk on the data store datastore2 {{(pid=61006) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 524.050928] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cd38f2-db0c-4494-9159-b7f151d6bf0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.061218] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e41d06a-134c-4bce-bdc0-8ffb7a91142a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.071207] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3895687-5d2d-4869-99a7-70137476241c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.107675] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44c5e8a-e08c-4624-86f7-f2236359c223 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.114547] env[61006]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-fca27d7d-bf31-4dff-adc9-6fe938a6c877 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.148036] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Downloading image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to the data store datastore2 {{(pid=61006) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 524.158783] env[61006]: DEBUG oslo_concurrency.lockutils [req-c44bc703-6d92-4373-a23c-02d8232e22c5 req-a95c3422-1f4a-44dd-b2bf-ebfa8f530ce5 service nova] Releasing lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.167376] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.192235] env[61006]: INFO nova.compute.manager [-] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Took 1.05 seconds to deallocate network for instance. [ 524.194235] env[61006]: DEBUG nova.compute.claims [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 524.194540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.239187] env[61006]: DEBUG oslo_vmware.rw_handles [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 524.439274] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.472057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquiring lock "f3a466c4-3b75-4f0c-8c17-076ccd1e258e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.473323] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "f3a466c4-3b75-4f0c-8c17-076ccd1e258e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.520232] env[61006]: DEBUG nova.scheduler.client.report [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.525402] env[61006]: DEBUG oslo_concurrency.lockutils [req-56d35dfd-99d2-40aa-8f80-7aaa6c94b0f8 req-451047b9-d152-44af-8433-7d132af6704f service nova] Releasing lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.611132] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.650017] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.650273] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.650428] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.650604] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.650742] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.652690] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.652690] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.652836] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.652999] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.653180] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.653357] env[61006]: DEBUG nova.virt.hardware [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.654265] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc38684-b224-4963-8460-67bbb68bf37b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.670025] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20df0ffb-06e2-42ab-a8a5-07ff3a384638 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.947848] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Releasing lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.949395] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.949395] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.949395] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45a5d4af-1a73-4b6c-bddd-1623b424ec16 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.962684] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1107723e-6ee9-4012-a19c-9502d31f9fcd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.977388] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.995765] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3769d396-ef6a-436b-ab75-faf66e4da813 could not be found. [ 524.995900] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.998304] env[61006]: INFO nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Took 0.05 seconds to destroy the instance on the hypervisor. [ 524.998304] env[61006]: DEBUG oslo.service.loopingcall [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.998304] env[61006]: DEBUG nova.compute.manager [-] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.998304] env[61006]: DEBUG nova.network.neutron [-] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 525.026437] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.026986] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.031134] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.929s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.031134] env[61006]: INFO nova.compute.claims [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.043187] env[61006]: DEBUG nova.network.neutron [-] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.122350] env[61006]: DEBUG oslo_vmware.rw_handles [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 525.122885] env[61006]: DEBUG oslo_vmware.rw_handles [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 525.170253] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Downloaded image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk on the data store datastore2 {{(pid=61006) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 525.172883] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 525.172883] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copying Virtual Disk [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk to [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 525.173670] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d05fe5c9-affd-4577-b54f-2cf5d96cee21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.184127] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 525.184127] env[61006]: value = "task-1336811" [ 525.184127] env[61006]: _type = "Task" [ 525.184127] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.201696] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336811, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.515582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.536347] env[61006]: DEBUG nova.compute.utils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.539609] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.539843] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.546173] env[61006]: DEBUG nova.network.neutron [-] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.630909] env[61006]: DEBUG nova.policy [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82aea0ad3a245af99868fc53a0b04f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd26132a67fc1412c938a36094a34a433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.696617] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336811, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.866436] env[61006]: DEBUG nova.compute.manager [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Received event network-vif-deleted-c55b3184-7509-4a17-bd6d-de59ec38bcba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.866657] env[61006]: DEBUG nova.compute.manager [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Received event network-changed-a24e2295-0c85-4318-a371-94d32be9aada {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.869218] env[61006]: DEBUG nova.compute.manager [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Refreshing instance network info cache due to event network-changed-a24e2295-0c85-4318-a371-94d32be9aada. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.869315] env[61006]: DEBUG oslo_concurrency.lockutils [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] Acquiring lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.869464] env[61006]: DEBUG oslo_concurrency.lockutils [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] Acquired lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.869625] env[61006]: DEBUG nova.network.neutron [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Refreshing network info cache for port a24e2295-0c85-4318-a371-94d32be9aada {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.913938] env[61006]: ERROR nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 525.913938] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.913938] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.913938] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.913938] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.913938] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.913938] env[61006]: ERROR nova.compute.manager raise self.value [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.913938] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 525.913938] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.913938] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 525.914386] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.914386] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 525.914386] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 525.914386] env[61006]: ERROR nova.compute.manager [ 525.914386] env[61006]: Traceback (most recent call last): [ 525.914386] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 525.914386] env[61006]: listener.cb(fileno) [ 525.914386] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.914386] env[61006]: result = function(*args, **kwargs) [ 525.914386] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.914386] env[61006]: return func(*args, **kwargs) [ 525.914386] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.914386] env[61006]: raise e [ 525.914386] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.914386] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 525.914386] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.914386] env[61006]: created_port_ids = self._update_ports_for_instance( [ 525.914386] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.914386] env[61006]: with excutils.save_and_reraise_exception(): [ 525.914386] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.914386] env[61006]: self.force_reraise() [ 525.914386] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.914386] env[61006]: raise self.value [ 525.914386] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.914386] env[61006]: updated_port = self._update_port( [ 525.914386] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.914386] env[61006]: _ensure_no_port_binding_failure(port) [ 525.914386] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.914386] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 525.915395] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 525.915395] env[61006]: Removing descriptor: 16 [ 525.915395] env[61006]: ERROR nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Traceback (most recent call last): [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] yield resources [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.driver.spawn(context, instance, image_meta, [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.915395] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] vm_ref = self.build_virtual_machine(instance, [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] for vif in network_info: [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self._sync_wrapper(fn, *args, **kwargs) [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.wait() [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self[:] = self._gt.wait() [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self._exit_event.wait() [ 525.915685] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] result = hub.switch() [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self.greenlet.switch() [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] result = function(*args, **kwargs) [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return func(*args, **kwargs) [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise e [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] nwinfo = self.network_api.allocate_for_instance( [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.916162] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] created_port_ids = self._update_ports_for_instance( [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] with excutils.save_and_reraise_exception(): [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.force_reraise() [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise self.value [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] updated_port = self._update_port( [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] _ensure_no_port_binding_failure(port) [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.916626] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise exception.PortBindingFailed(port_id=port['id']) [ 525.916906] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 525.916906] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] [ 525.916906] env[61006]: INFO nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Terminating instance [ 525.919783] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.920076] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquired lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.920303] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.044706] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.056961] env[61006]: INFO nova.compute.manager [-] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Took 1.06 seconds to deallocate network for instance. [ 526.063896] env[61006]: DEBUG nova.compute.claims [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.064104] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.084361] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Successfully created port: 9759c4cb-2acb-431c-aa62-4718e447227b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.172274] env[61006]: DEBUG nova.compute.manager [req-c66c1849-05cb-4adf-8a3f-a32fc0316852 req-201dbb49-172f-4121-be87-23a0cfc9ce54 service nova] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Received event network-vif-deleted-03ae18ba-2881-4fe6-8eb1-5c6066c15be4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.199596] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336811, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726596} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.199859] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copied Virtual Disk [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk to [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 526.200065] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleting the datastore file [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 526.200326] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9080f255-d6da-4aa1-8cfa-d245102043ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.208555] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 526.208555] env[61006]: value = "task-1336812" [ 526.208555] env[61006]: _type = "Task" [ 526.208555] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.220914] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.298375] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4122981-9f76-475c-9c0d-913be85416b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.305297] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71a6129-8115-415b-9374-e3979864579a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.340810] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a075263e-77d0-4d00-aada-5fa476d3edf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.349505] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e7cee2-5055-4653-b2ad-f4b16694dab3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.365061] env[61006]: DEBUG nova.compute.provider_tree [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.409193] env[61006]: DEBUG nova.network.neutron [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.459482] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.595410] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.685311] env[61006]: DEBUG nova.network.neutron [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.720500] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024671} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.720722] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 526.720922] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Moving file from [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0. {{(pid=61006) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 526.721187] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a81e6414-a23e-45f1-bf1b-730c8a43cab8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.729090] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 526.729090] env[61006]: value = "task-1336813" [ 526.729090] env[61006]: _type = "Task" [ 526.729090] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.738999] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336813, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.868350] env[61006]: DEBUG nova.scheduler.client.report [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.065527] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.098070] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Releasing lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.098962] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.098962] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.099272] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e163ea59-e9e3-4736-a646-49e7726c9464 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.103729] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.104066] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.104162] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.104316] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.104508] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.104714] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.105084] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.105084] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.105148] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.105345] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.105472] env[61006]: DEBUG nova.virt.hardware [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.106312] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062a23f9-e3ab-4705-b29e-a629bcb33753 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.119092] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7067d1-efbd-4f26-a245-4e252ab67ff7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.137312] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c09118a-5f5e-4073-b62b-1d411bc73615 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.155212] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06d7a63f-e04b-49d0-8827-ef25a7c9d11d could not be found. [ 527.155212] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.155212] env[61006]: INFO nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 527.155212] env[61006]: DEBUG oslo.service.loopingcall [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.155212] env[61006]: DEBUG nova.compute.manager [-] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.155212] env[61006]: DEBUG nova.network.neutron [-] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.190056] env[61006]: DEBUG oslo_concurrency.lockutils [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] Releasing lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.191275] env[61006]: DEBUG nova.compute.manager [req-e35003c7-66f1-4fb6-ae67-828190fef0cf req-d967af2d-ff77-43fa-bff7-4eaeb4eb658c service nova] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Received event network-vif-deleted-a24e2295-0c85-4318-a371-94d32be9aada {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.209313] env[61006]: DEBUG nova.network.neutron [-] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.240187] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336813, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.039546} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.240455] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] File moved {{(pid=61006) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 527.240639] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Cleaning up location [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 527.240821] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleting the datastore file [datastore2] vmware_temp/a69db071-9001-4d8b-95d6-5ea82975bc7d {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 527.241297] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a51dc11-8cca-4639-b784-fe524e99c8df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.249195] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 527.249195] env[61006]: value = "task-1336814" [ 527.249195] env[61006]: _type = "Task" [ 527.249195] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.261566] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.375069] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.375069] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.382292] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.251s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.383776] env[61006]: INFO nova.compute.claims [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.713910] env[61006]: DEBUG nova.network.neutron [-] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.761731] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336814, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026905} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.762027] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 527.762763] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b21598c0-bac7-44ed-a56d-de07726a7cd8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.768355] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 527.768355] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dcf4f2-0bf4-0a67-9761-7517f13c14d3" [ 527.768355] env[61006]: _type = "Task" [ 527.768355] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.777713] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dcf4f2-0bf4-0a67-9761-7517f13c14d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.889554] env[61006]: DEBUG nova.compute.utils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.893544] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.893753] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 527.979372] env[61006]: DEBUG nova.policy [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab0648a37b8a4aaba937edced69c20f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86d059b3ce274cb7a16c9e4bb4f843c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.216143] env[61006]: INFO nova.compute.manager [-] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Took 1.06 seconds to deallocate network for instance. [ 528.219655] env[61006]: DEBUG nova.compute.claims [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.219875] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.224108] env[61006]: ERROR nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 528.224108] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.224108] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.224108] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.224108] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.224108] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.224108] env[61006]: ERROR nova.compute.manager raise self.value [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.224108] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.224108] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.224108] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.224512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.224512] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.224512] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 528.224512] env[61006]: ERROR nova.compute.manager [ 528.224512] env[61006]: Traceback (most recent call last): [ 528.224512] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.224512] env[61006]: listener.cb(fileno) [ 528.224512] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.224512] env[61006]: result = function(*args, **kwargs) [ 528.224512] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 528.224512] env[61006]: return func(*args, **kwargs) [ 528.224512] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.224512] env[61006]: raise e [ 528.224512] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.224512] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 528.224512] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.224512] env[61006]: created_port_ids = self._update_ports_for_instance( [ 528.224512] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.224512] env[61006]: with excutils.save_and_reraise_exception(): [ 528.224512] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.224512] env[61006]: self.force_reraise() [ 528.224512] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.224512] env[61006]: raise self.value [ 528.224512] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.224512] env[61006]: updated_port = self._update_port( [ 528.224512] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.224512] env[61006]: _ensure_no_port_binding_failure(port) [ 528.224512] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.224512] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.225207] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 528.225207] env[61006]: Removing descriptor: 15 [ 528.225207] env[61006]: ERROR nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Traceback (most recent call last): [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] yield resources [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.driver.spawn(context, instance, image_meta, [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.225207] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] vm_ref = self.build_virtual_machine(instance, [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] for vif in network_info: [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self._sync_wrapper(fn, *args, **kwargs) [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.wait() [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self[:] = self._gt.wait() [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self._exit_event.wait() [ 528.225550] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] result = hub.switch() [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self.greenlet.switch() [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] result = function(*args, **kwargs) [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return func(*args, **kwargs) [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise e [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] nwinfo = self.network_api.allocate_for_instance( [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.225919] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] created_port_ids = self._update_ports_for_instance( [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] with excutils.save_and_reraise_exception(): [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.force_reraise() [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise self.value [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] updated_port = self._update_port( [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] _ensure_no_port_binding_failure(port) [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.226381] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise exception.PortBindingFailed(port_id=port['id']) [ 528.226684] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 528.226684] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] [ 528.226684] env[61006]: INFO nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Terminating instance [ 528.227843] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.228094] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.229057] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 528.292111] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dcf4f2-0bf4-0a67-9761-7517f13c14d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010264} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.292111] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.292111] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 528.292111] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94fffc61-852c-4aa9-b510-4699721ea755 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.303013] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 528.303013] env[61006]: value = "task-1336815" [ 528.303013] env[61006]: _type = "Task" [ 528.303013] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.313560] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336815, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.395506] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.590591] env[61006]: DEBUG nova.compute.manager [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Received event network-changed-9759c4cb-2acb-431c-aa62-4718e447227b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 528.590874] env[61006]: DEBUG nova.compute.manager [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Refreshing instance network info cache due to event network-changed-9759c4cb-2acb-431c-aa62-4718e447227b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 528.591037] env[61006]: DEBUG oslo_concurrency.lockutils [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] Acquiring lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.618083] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "116d50e2-d8fa-4e26-8301-d1066d627982" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.618083] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.628838] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a72dbf-02f3-4014-b6ec-b296b0abd64a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.638653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84bb1bf-2bcd-4ee0-8145-e7ed92cb3ddf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.678433] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb085ab-83c1-4552-a5e1-e1641b47caae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.686325] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4168e603-7a14-4f4c-bc6b-2a5135815345 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.700613] env[61006]: DEBUG nova.compute.provider_tree [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.765528] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.814573] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336815, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.878934] env[61006]: DEBUG nova.compute.manager [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Received event network-changed-3518a742-741e-4531-9e63-a8e3e8d13f1e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 528.879145] env[61006]: DEBUG nova.compute.manager [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Refreshing instance network info cache due to event network-changed-3518a742-741e-4531-9e63-a8e3e8d13f1e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 528.879342] env[61006]: DEBUG oslo_concurrency.lockutils [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] Acquiring lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.879471] env[61006]: DEBUG oslo_concurrency.lockutils [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] Acquired lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.879636] env[61006]: DEBUG nova.network.neutron [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Refreshing network info cache for port 3518a742-741e-4531-9e63-a8e3e8d13f1e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 528.939602] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.059032] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Successfully created port: 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.206512] env[61006]: DEBUG nova.scheduler.client.report [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.317746] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336815, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532653} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.318226] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 529.319050] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 529.319050] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e982543-396a-4f2d-accf-d7f67c02ad2e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.328324] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 529.328324] env[61006]: value = "task-1336816" [ 529.328324] env[61006]: _type = "Task" [ 529.328324] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.337347] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336816, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.409168] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.421347] env[61006]: DEBUG nova.network.neutron [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.447810] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.448408] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.449317] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.449568] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.451361] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.451361] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.451361] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.451361] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.451361] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.451550] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.451550] env[61006]: DEBUG nova.virt.hardware [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.452415] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.452722] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.452990] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.453925] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c623e30-e8f6-416c-a33d-c04c46a0cf75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.460488] env[61006]: DEBUG oslo_concurrency.lockutils [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] Acquired lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.461264] env[61006]: DEBUG nova.network.neutron [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Refreshing network info cache for port 9759c4cb-2acb-431c-aa62-4718e447227b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 529.462806] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46e17a5f-5d95-4493-82ea-814d7c11d3bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.471440] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879ad2c4-b6ae-41a4-87a7-025613fa910e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.479224] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7231aef4-76d8-45c6-b2e6-c7f6d4b35a37 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.507561] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94 could not be found. [ 529.507840] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.508124] env[61006]: INFO nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Took 0.06 seconds to destroy the instance on the hypervisor. [ 529.508418] env[61006]: DEBUG oslo.service.loopingcall [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.508669] env[61006]: DEBUG nova.compute.manager [-] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.508793] env[61006]: DEBUG nova.network.neutron [-] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.548441] env[61006]: DEBUG nova.network.neutron [-] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.604253] env[61006]: DEBUG nova.network.neutron [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.715709] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.716363] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.719382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.170s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.723456] env[61006]: INFO nova.compute.claims [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.838723] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336816, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071618} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.839420] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 529.840222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4dd46e-1964-4be9-ba27-70d2cfe119a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.867874] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 529.867874] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23165d7c-97ee-478a-bef5-1070e335e1e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.889922] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 529.889922] env[61006]: value = "task-1336817" [ 529.889922] env[61006]: _type = "Task" [ 529.889922] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.900210] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336817, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.009256] env[61006]: DEBUG nova.network.neutron [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.055974] env[61006]: DEBUG nova.network.neutron [-] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.110611] env[61006]: DEBUG oslo_concurrency.lockutils [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] Releasing lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.111046] env[61006]: DEBUG nova.compute.manager [req-084dbcbd-681c-4bd9-a05d-6b0dbf893ab0 req-4a346f2e-8407-49e3-8d0e-9b074d1a8d70 service nova] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Received event network-vif-deleted-3518a742-741e-4531-9e63-a8e3e8d13f1e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.227294] env[61006]: DEBUG nova.compute.utils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.234939] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.235073] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 530.402096] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336817, 'name': ReconfigVM_Task, 'duration_secs': 0.285613} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.402397] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfigured VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 530.403077] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85c5065f-efc0-4737-8e6a-648bbb98249d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.408942] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 530.408942] env[61006]: value = "task-1336818" [ 530.408942] env[61006]: _type = "Task" [ 530.408942] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.421579] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336818, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.427543] env[61006]: DEBUG nova.network.neutron [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.529878] env[61006]: DEBUG nova.policy [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6ec4fc65004492913a32b5e0afdfdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6a377db9cf846fcbe371d3ab34dc972', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.559928] env[61006]: INFO nova.compute.manager [-] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Took 1.05 seconds to deallocate network for instance. [ 530.563439] env[61006]: DEBUG nova.compute.claims [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.563526] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.741343] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.919445] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336818, 'name': Rename_Task, 'duration_secs': 0.152767} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.921954] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 530.922366] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e53ab19-6ffb-403c-ada3-1b0fb0eef026 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.928772] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 530.928772] env[61006]: value = "task-1336819" [ 530.928772] env[61006]: _type = "Task" [ 530.928772] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.933522] env[61006]: DEBUG oslo_concurrency.lockutils [req-16d366f1-6530-4577-a98b-3effc49a3b71 req-2480d865-5395-4714-ade9-674fda31f769 service nova] Releasing lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.934682] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c0387a-273f-4e4f-bfa7-6fc5bcf122bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.942463] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.948825] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16861d0-09e4-4a8a-9bf2-9f4c46ae8690 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.980281] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.982048] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.982048] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 530.982048] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 530.983652] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e7e7a4-4e40-4e52-95cd-ffd6eb0ada59 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.991560] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f615a17d-4ba7-4002-a7fa-1cc1cd17ce30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.006115] env[61006]: DEBUG nova.compute.provider_tree [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.442336] env[61006]: DEBUG oslo_vmware.api [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336819, 'name': PowerOnVM_Task, 'duration_secs': 0.42457} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.442626] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 531.442970] env[61006]: INFO nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Took 9.17 seconds to spawn the instance on the hypervisor. [ 531.443092] env[61006]: DEBUG nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.444965] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6decd7-bb47-4eef-9a57-9fa83fe74bdb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.488685] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.488685] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.488685] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.489875] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.489875] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.489875] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.489875] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.489875] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.492172] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 531.495794] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Didn't find any instances for network info cache update. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 531.496035] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.496998] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.496998] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.496998] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.496998] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.496998] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.509341] env[61006]: DEBUG nova.scheduler.client.report [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.754279] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.797450] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.797450] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.797450] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.797801] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.797801] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.797801] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.800050] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.800772] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.803615] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.803615] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.803615] env[61006]: DEBUG nova.virt.hardware [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.804237] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710e887d-03a9-49a9-a513-9e7210bfabee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.817302] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef130f16-cf02-4131-b4ae-653b81a576ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.967869] env[61006]: INFO nova.compute.manager [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Took 16.11 seconds to build instance. [ 531.970516] env[61006]: DEBUG nova.compute.manager [req-94ebb019-2341-4509-aeb9-90cfc5ba44f8 req-cf1152b9-390f-4949-876e-8ab557248370 service nova] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Received event network-vif-deleted-9759c4cb-2acb-431c-aa62-4718e447227b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 532.002032] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Getting list of instances from cluster (obj){ [ 532.002032] env[61006]: value = "domain-c8" [ 532.002032] env[61006]: _type = "ClusterComputeResource" [ 532.002032] env[61006]: } {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 532.002792] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a48e018-fdee-4ab6-a72f-6ceb2557dda9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.016386] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.017052] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.020095] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Got total of 1 instances {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 532.020253] env[61006]: WARNING nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] While synchronizing instance power states, found 9 instances in the database and 1 instances on the hypervisor. [ 532.020461] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid f3feb305-2737-4398-84d3-311a4f9c42fc {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.020645] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid a4d93990-9591-4bb5-8e26-d8a57807d8d0 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.020803] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 3769d396-ef6a-436b-ab75-faf66e4da813 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.020985] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid a2360f64-2bed-4c0a-9f99-54b2a34f8d68 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.021208] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 06d7a63f-e04b-49d0-8827-ef25a7c9d11d {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.021391] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.021545] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 0ace1533-286c-42e2-afde-e99c77fdd747 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.021687] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 583945d7-2f35-48b1-9bb7-26358a2b0d2f {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.021878] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid fa649a9c-8745-4188-80b8-09d62cd882e5 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 532.022237] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.445s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.025973] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "f3feb305-2737-4398-84d3-311a4f9c42fc" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.030103] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.030415] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "3769d396-ef6a-436b-ab75-faf66e4da813" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031288] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "0ace1533-286c-42e2-afde-e99c77fdd747" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031698] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "fa649a9c-8745-4188-80b8-09d62cd882e5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.031878] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.032107] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 532.033089] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 532.170858] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Successfully created port: 7046e781-c7d6-4859-81be-ff7f55f4856b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.471669] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8060f11c-391c-4377-b085-73bf4afb9110 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.629s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.472972] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.442s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.474068] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] During sync_power_state the instance has a pending task (spawning). Skip. [ 532.474068] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.535493] env[61006]: DEBUG nova.compute.utils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.543613] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.543953] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 532.547435] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.770731] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52abc24f-e872-4df8-8742-f12b53741551 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.779300] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d284a9c3-de13-4a90-a8d0-3b24c786ca74 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.814952] env[61006]: DEBUG nova.policy [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93d3f89116184ea5a2b119c4c6fb59f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '590072e79fac4e969ffc6692874ebd4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.817077] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf5e629-0e39-48f6-8acb-e83f531727f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.825565] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda134ff-ff97-4f54-8ed9-58dedbff7d53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.839947] env[61006]: DEBUG nova.compute.provider_tree [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.976959] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.044923] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.344176] env[61006]: DEBUG nova.scheduler.client.report [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.510462] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.851420] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.829s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.851774] env[61006]: ERROR nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Traceback (most recent call last): [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.driver.spawn(context, instance, image_meta, [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] vm_ref = self.build_virtual_machine(instance, [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.851774] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] for vif in network_info: [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self._sync_wrapper(fn, *args, **kwargs) [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.wait() [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self[:] = self._gt.wait() [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self._exit_event.wait() [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] result = hub.switch() [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.852481] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return self.greenlet.switch() [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] result = function(*args, **kwargs) [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] return func(*args, **kwargs) [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise e [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] nwinfo = self.network_api.allocate_for_instance( [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] created_port_ids = self._update_ports_for_instance( [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] with excutils.save_and_reraise_exception(): [ 533.853046] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] self.force_reraise() [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise self.value [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] updated_port = self._update_port( [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] _ensure_no_port_binding_failure(port) [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] raise exception.PortBindingFailed(port_id=port['id']) [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] nova.exception.PortBindingFailed: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. [ 533.853579] env[61006]: ERROR nova.compute.manager [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] [ 533.854051] env[61006]: DEBUG nova.compute.utils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 533.855800] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.661s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.859794] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Build of instance f3feb305-2737-4398-84d3-311a4f9c42fc was re-scheduled: Binding failed for port c55b3184-7509-4a17-bd6d-de59ec38bcba, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 533.859862] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 533.860138] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquiring lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.860234] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Acquired lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.860420] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.056899] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.094493] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.094719] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.094939] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.095204] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.095384] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.095583] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.095844] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.096056] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.096270] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.096506] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.096658] env[61006]: DEBUG nova.virt.hardware [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.097619] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0c1168-4c48-40bc-a769-1ac4e00bdab6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.114354] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70d4662-7a10-4d9f-94fe-92884b7f08a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.489285] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Successfully created port: 3dd5dc5a-191d-4c53-a527-b28e518fbf2a {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.520501] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.580352] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316db11a-21ef-4448-8916-03333169b766 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.588517] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d519bc05-14c1-4e94-85b8-dce7b3e2b931 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.623946] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a5dd54-d69c-44e1-8cd4-27861b1cf4bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.632480] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db76f829-6e80-46e1-90d7-43e0ca13940b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.647567] env[61006]: DEBUG nova.compute.provider_tree [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.819705] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquiring lock "488fc6b9-9398-4a8b-a471-0ba7323c9f9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.819933] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "488fc6b9-9398-4a8b-a471-0ba7323c9f9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.824945] env[61006]: ERROR nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 534.824945] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.824945] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.824945] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.824945] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.824945] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.824945] env[61006]: ERROR nova.compute.manager raise self.value [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.824945] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.824945] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.824945] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.825412] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.825412] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.825412] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 534.825412] env[61006]: ERROR nova.compute.manager [ 534.825412] env[61006]: Traceback (most recent call last): [ 534.825412] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.825412] env[61006]: listener.cb(fileno) [ 534.825412] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.825412] env[61006]: result = function(*args, **kwargs) [ 534.825412] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.825412] env[61006]: return func(*args, **kwargs) [ 534.825412] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.825412] env[61006]: raise e [ 534.825412] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.825412] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 534.825412] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.825412] env[61006]: created_port_ids = self._update_ports_for_instance( [ 534.825412] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.825412] env[61006]: with excutils.save_and_reraise_exception(): [ 534.825412] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.825412] env[61006]: self.force_reraise() [ 534.825412] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.825412] env[61006]: raise self.value [ 534.825412] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.825412] env[61006]: updated_port = self._update_port( [ 534.825412] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.825412] env[61006]: _ensure_no_port_binding_failure(port) [ 534.825412] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.825412] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.826842] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 534.826842] env[61006]: Removing descriptor: 16 [ 534.826842] env[61006]: ERROR nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Traceback (most recent call last): [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] yield resources [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.driver.spawn(context, instance, image_meta, [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.826842] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] vm_ref = self.build_virtual_machine(instance, [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] for vif in network_info: [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self._sync_wrapper(fn, *args, **kwargs) [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.wait() [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self[:] = self._gt.wait() [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self._exit_event.wait() [ 534.827189] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] result = hub.switch() [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self.greenlet.switch() [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] result = function(*args, **kwargs) [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return func(*args, **kwargs) [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise e [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] nwinfo = self.network_api.allocate_for_instance( [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.827523] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] created_port_ids = self._update_ports_for_instance( [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] with excutils.save_and_reraise_exception(): [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.force_reraise() [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise self.value [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] updated_port = self._update_port( [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] _ensure_no_port_binding_failure(port) [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.827855] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise exception.PortBindingFailed(port_id=port['id']) [ 534.828211] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 534.828211] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] [ 534.828211] env[61006]: INFO nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Terminating instance [ 534.828211] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.828326] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquired lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.828607] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.928186] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.151190] env[61006]: DEBUG nova.scheduler.client.report [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.432883] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Releasing lock "refresh_cache-f3feb305-2737-4398-84d3-311a4f9c42fc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.433135] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 535.433298] env[61006]: DEBUG nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.433506] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.470131] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.651098] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.657167] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.801s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.657851] env[61006]: ERROR nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Traceback (most recent call last): [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.driver.spawn(context, instance, image_meta, [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] vm_ref = self.build_virtual_machine(instance, [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.657851] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] for vif in network_info: [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self._sync_wrapper(fn, *args, **kwargs) [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.wait() [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self[:] = self._gt.wait() [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self._exit_event.wait() [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] result = hub.switch() [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.658257] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return self.greenlet.switch() [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] result = function(*args, **kwargs) [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] return func(*args, **kwargs) [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise e [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] nwinfo = self.network_api.allocate_for_instance( [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] created_port_ids = self._update_ports_for_instance( [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] with excutils.save_and_reraise_exception(): [ 535.658576] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] self.force_reraise() [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise self.value [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] updated_port = self._update_port( [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] _ensure_no_port_binding_failure(port) [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] raise exception.PortBindingFailed(port_id=port['id']) [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] nova.exception.PortBindingFailed: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. [ 535.658884] env[61006]: ERROR nova.compute.manager [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] [ 535.659231] env[61006]: DEBUG nova.compute.utils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 535.661226] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Build of instance a4d93990-9591-4bb5-8e26-d8a57807d8d0 was re-scheduled: Binding failed for port 03ae18ba-2881-4fe6-8eb1-5c6066c15be4, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.661708] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.662047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.662207] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.662429] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 535.666083] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.151s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.667854] env[61006]: INFO nova.compute.claims [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.710107] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.155306] env[61006]: DEBUG nova.network.neutron [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.212949] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Releasing lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.213645] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.213645] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.213900] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e4e3536-e463-4958-900d-a7a559f86066 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.219701] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.229227] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a27cc0-7ec4-4658-9426-b8117dbfeb00 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.254511] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ace1533-286c-42e2-afde-e99c77fdd747 could not be found. [ 536.254761] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.254947] env[61006]: INFO nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.256246] env[61006]: DEBUG oslo.service.loopingcall [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.256679] env[61006]: DEBUG nova.compute.manager [-] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.256755] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 536.271943] env[61006]: DEBUG nova.compute.manager [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Received event network-changed-1181ed42-1c4d-4ea7-a4b9-437f44a01bdc {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.272146] env[61006]: DEBUG nova.compute.manager [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Refreshing instance network info cache due to event network-changed-1181ed42-1c4d-4ea7-a4b9-437f44a01bdc. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 536.273430] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] Acquiring lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.273430] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] Acquired lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.273430] env[61006]: DEBUG nova.network.neutron [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Refreshing network info cache for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 536.342905] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.535666] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.663583] env[61006]: INFO nova.compute.manager [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] Took 1.23 seconds to deallocate network for instance. [ 536.834337] env[61006]: DEBUG nova.network.neutron [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.847238] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.899299] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquiring lock "25ebb89a-1054-4447-83fc-91c59e4fc80b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.899615] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "25ebb89a-1054-4447-83fc-91c59e4fc80b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.919424] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "41826fe1-8f39-479a-b9fd-51399753dfb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.919641] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.961986] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bca23aa-1f8c-41f8-8ded-b5c044775be1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.972150] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a883fc-21f8-4612-b976-3be6084b78c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.008970] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cfa7a5-ca10-4e51-9ec4-ecccc7f45043 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.017566] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff83e62-9836-4818-9a77-687b29ef85bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.035544] env[61006]: DEBUG nova.compute.provider_tree [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.037629] env[61006]: DEBUG nova.network.neutron [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.038951] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-a4d93990-9591-4bb5-8e26-d8a57807d8d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.039119] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 537.039356] env[61006]: DEBUG nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.039571] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.087994] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.317827] env[61006]: INFO nova.compute.manager [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Rebuilding instance [ 537.349974] env[61006]: INFO nova.compute.manager [-] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Took 1.09 seconds to deallocate network for instance. [ 537.354196] env[61006]: DEBUG nova.compute.claims [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.354619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.380794] env[61006]: DEBUG nova.compute.manager [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.381554] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b2d326-dd3a-4ad6-a48d-25e6ad5ec860 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.543244] env[61006]: DEBUG nova.scheduler.client.report [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.551884] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef5b017d-dd86-45b5-b4ca-6df698cd0bea req-c25dc090-f3a3-4dbb-b43d-4cbced4645ee service nova] Releasing lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.592775] env[61006]: DEBUG nova.network.neutron [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.695885] env[61006]: INFO nova.scheduler.client.report [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Deleted allocations for instance f3feb305-2737-4398-84d3-311a4f9c42fc [ 537.821729] env[61006]: ERROR nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 537.821729] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.821729] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.821729] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.821729] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.821729] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.821729] env[61006]: ERROR nova.compute.manager raise self.value [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.821729] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.821729] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.821729] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.822193] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.822193] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.822193] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 537.822193] env[61006]: ERROR nova.compute.manager [ 537.822193] env[61006]: Traceback (most recent call last): [ 537.822193] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.822193] env[61006]: listener.cb(fileno) [ 537.822193] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.822193] env[61006]: result = function(*args, **kwargs) [ 537.822193] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.822193] env[61006]: return func(*args, **kwargs) [ 537.822193] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.822193] env[61006]: raise e [ 537.822193] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.822193] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 537.822193] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.822193] env[61006]: created_port_ids = self._update_ports_for_instance( [ 537.822193] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.822193] env[61006]: with excutils.save_and_reraise_exception(): [ 537.822193] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.822193] env[61006]: self.force_reraise() [ 537.822193] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.822193] env[61006]: raise self.value [ 537.822193] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.822193] env[61006]: updated_port = self._update_port( [ 537.822193] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.822193] env[61006]: _ensure_no_port_binding_failure(port) [ 537.822193] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.822193] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.822967] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 537.822967] env[61006]: Removing descriptor: 17 [ 537.822967] env[61006]: ERROR nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Traceback (most recent call last): [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] yield resources [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.driver.spawn(context, instance, image_meta, [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.822967] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] vm_ref = self.build_virtual_machine(instance, [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] for vif in network_info: [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self._sync_wrapper(fn, *args, **kwargs) [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.wait() [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self[:] = self._gt.wait() [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self._exit_event.wait() [ 537.823396] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] result = hub.switch() [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self.greenlet.switch() [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] result = function(*args, **kwargs) [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return func(*args, **kwargs) [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise e [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] nwinfo = self.network_api.allocate_for_instance( [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.823742] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] created_port_ids = self._update_ports_for_instance( [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] with excutils.save_and_reraise_exception(): [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.force_reraise() [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise self.value [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] updated_port = self._update_port( [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] _ensure_no_port_binding_failure(port) [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.824342] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise exception.PortBindingFailed(port_id=port['id']) [ 537.824669] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 537.824669] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] [ 537.824669] env[61006]: INFO nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Terminating instance [ 537.826733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquiring lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.826886] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquired lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.827788] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.893878] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 537.895214] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4ed6744-e642-4c5b-900e-48b1d69b04ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.904859] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 537.904859] env[61006]: value = "task-1336820" [ 537.904859] env[61006]: _type = "Task" [ 537.904859] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.917152] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.055210] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.055484] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.059670] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.995s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.100771] env[61006]: INFO nova.compute.manager [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] Took 1.06 seconds to deallocate network for instance. [ 538.210745] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdde486f-07f3-4afd-a811-a701ad5bf28e tempest-ServersAdminNegativeTestJSON-2048520182 tempest-ServersAdminNegativeTestJSON-2048520182-project-member] Lock "f3feb305-2737-4398-84d3-311a4f9c42fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.977s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.212115] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "f3feb305-2737-4398-84d3-311a4f9c42fc" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.186s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.213960] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: f3feb305-2737-4398-84d3-311a4f9c42fc] During sync_power_state the instance has a pending task (spawning). Skip. [ 538.213960] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "f3feb305-2737-4398-84d3-311a4f9c42fc" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.373964] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.420069] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336820, 'name': PowerOffVM_Task, 'duration_secs': 0.11208} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.420069] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 538.420069] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.420069] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a258af07-75fe-464e-a02d-98c082781b31 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.430957] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 538.431245] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10a9212d-4510-4ee6-8340-2e513ff191a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.453238] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 538.453454] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 538.453632] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleting the datastore file [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 538.453890] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f2b2186-0a69-478b-9768-359d6547bb32 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.464793] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 538.464793] env[61006]: value = "task-1336822" [ 538.464793] env[61006]: _type = "Task" [ 538.464793] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.480437] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.568073] env[61006]: DEBUG nova.compute.utils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.574109] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.574109] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 538.635930] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.713702] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.805028] env[61006]: DEBUG nova.policy [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4966179da7ca4222a922da6b5ec9588a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c79a4b9437d24a4898273c90c238b3dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.825232] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c29e3a-f64a-4044-a172-5781ee8f361d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.834545] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd11bc7-3181-41e6-a0a7-6853bdc0ff14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.870397] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f024b2a2-8457-47b6-9bd7-89658e2118f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.880825] env[61006]: DEBUG nova.compute.manager [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Received event network-changed-7046e781-c7d6-4859-81be-ff7f55f4856b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.880825] env[61006]: DEBUG nova.compute.manager [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Refreshing instance network info cache due to event network-changed-7046e781-c7d6-4859-81be-ff7f55f4856b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.880825] env[61006]: DEBUG oslo_concurrency.lockutils [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] Acquiring lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.886033] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91b68ec-6901-4329-aad2-f8e632b53c1e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.903848] env[61006]: DEBUG nova.compute.provider_tree [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.981610] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0966} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.982080] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 538.982146] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 538.982288] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.073981] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.140889] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Releasing lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.141378] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.141562] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.142223] env[61006]: DEBUG oslo_concurrency.lockutils [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] Acquired lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.142401] env[61006]: DEBUG nova.network.neutron [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Refreshing network info cache for port 7046e781-c7d6-4859-81be-ff7f55f4856b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 539.144137] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aad6e612-690b-447c-bb03-3f739676eaa0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.153021] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffe012f-72a2-4387-8958-6adbaf94b19c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.165857] env[61006]: INFO nova.scheduler.client.report [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance a4d93990-9591-4bb5-8e26-d8a57807d8d0 [ 539.184652] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 583945d7-2f35-48b1-9bb7-26358a2b0d2f could not be found. [ 539.184652] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.184761] env[61006]: INFO nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 539.184983] env[61006]: DEBUG oslo.service.loopingcall [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.185198] env[61006]: DEBUG nova.compute.manager [-] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.185285] env[61006]: DEBUG nova.network.neutron [-] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.240107] env[61006]: DEBUG nova.network.neutron [-] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.246468] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.279249] env[61006]: ERROR nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 539.279249] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.279249] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.279249] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.279249] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.279249] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.279249] env[61006]: ERROR nova.compute.manager raise self.value [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.279249] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.279249] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.279249] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.280096] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.280096] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.280096] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 539.280096] env[61006]: ERROR nova.compute.manager [ 539.280096] env[61006]: Traceback (most recent call last): [ 539.280096] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.280096] env[61006]: listener.cb(fileno) [ 539.280096] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.280096] env[61006]: result = function(*args, **kwargs) [ 539.280096] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.280096] env[61006]: return func(*args, **kwargs) [ 539.280096] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.280096] env[61006]: raise e [ 539.280096] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.280096] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 539.280096] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.280096] env[61006]: created_port_ids = self._update_ports_for_instance( [ 539.280096] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.280096] env[61006]: with excutils.save_and_reraise_exception(): [ 539.280096] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.280096] env[61006]: self.force_reraise() [ 539.280096] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.280096] env[61006]: raise self.value [ 539.280096] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.280096] env[61006]: updated_port = self._update_port( [ 539.280096] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.280096] env[61006]: _ensure_no_port_binding_failure(port) [ 539.280096] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.280096] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.280860] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 539.280860] env[61006]: Removing descriptor: 15 [ 539.280860] env[61006]: ERROR nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Traceback (most recent call last): [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] yield resources [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.driver.spawn(context, instance, image_meta, [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.280860] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] vm_ref = self.build_virtual_machine(instance, [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] for vif in network_info: [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self._sync_wrapper(fn, *args, **kwargs) [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.wait() [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self[:] = self._gt.wait() [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self._exit_event.wait() [ 539.281505] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] result = hub.switch() [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self.greenlet.switch() [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] result = function(*args, **kwargs) [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return func(*args, **kwargs) [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise e [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] nwinfo = self.network_api.allocate_for_instance( [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.282121] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] created_port_ids = self._update_ports_for_instance( [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] with excutils.save_and_reraise_exception(): [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.force_reraise() [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise self.value [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] updated_port = self._update_port( [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] _ensure_no_port_binding_failure(port) [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.282823] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise exception.PortBindingFailed(port_id=port['id']) [ 539.283337] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 539.283337] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] [ 539.283337] env[61006]: INFO nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Terminating instance [ 539.283691] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquiring lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.283881] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquired lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.284081] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.408116] env[61006]: DEBUG nova.scheduler.client.report [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.633920] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquiring lock "075e54ee-fc17-426d-ac3a-a632699274a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.633920] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "075e54ee-fc17-426d-ac3a-a632699274a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.679154] env[61006]: DEBUG oslo_concurrency.lockutils [None req-021b8c0b-9b50-4df3-8a30-fdd657a4fab0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.501s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.679360] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.649s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.679553] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a4d93990-9591-4bb5-8e26-d8a57807d8d0] During sync_power_state the instance has a pending task (spawning). Skip. [ 539.679732] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "a4d93990-9591-4bb5-8e26-d8a57807d8d0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.690409] env[61006]: DEBUG nova.network.neutron [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.703759] env[61006]: DEBUG nova.compute.manager [req-ff31e7bd-1c5d-4ab2-9da1-e727c0179d92 req-ee951efd-bad5-42aa-b1fa-bc3cadc298f9 service nova] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Received event network-vif-deleted-1181ed42-1c4d-4ea7-a4b9-437f44a01bdc {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.744144] env[61006]: DEBUG nova.network.neutron [-] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.853551] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.917818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.858s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.917982] env[61006]: ERROR nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Traceback (most recent call last): [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.driver.spawn(context, instance, image_meta, [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] vm_ref = self.build_virtual_machine(instance, [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.917982] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] for vif in network_info: [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self._sync_wrapper(fn, *args, **kwargs) [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.wait() [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self[:] = self._gt.wait() [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self._exit_event.wait() [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] result = hub.switch() [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.918445] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return self.greenlet.switch() [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] result = function(*args, **kwargs) [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] return func(*args, **kwargs) [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise e [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] nwinfo = self.network_api.allocate_for_instance( [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] created_port_ids = self._update_ports_for_instance( [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] with excutils.save_and_reraise_exception(): [ 539.918831] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] self.force_reraise() [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise self.value [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] updated_port = self._update_port( [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] _ensure_no_port_binding_failure(port) [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] raise exception.PortBindingFailed(port_id=port['id']) [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] nova.exception.PortBindingFailed: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. [ 539.919207] env[61006]: ERROR nova.compute.manager [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] [ 539.919524] env[61006]: DEBUG nova.compute.utils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 539.921598] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.701s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.926862] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Build of instance 3769d396-ef6a-436b-ab75-faf66e4da813 was re-scheduled: Binding failed for port a24e2295-0c85-4318-a371-94d32be9aada, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 539.926862] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 539.926862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquiring lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.926862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Acquired lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.927219] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.024752] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.025451] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.025451] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.025451] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.025451] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.025642] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.025790] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.025946] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.026651] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.026880] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.027073] env[61006]: DEBUG nova.virt.hardware [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.027945] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9d463b-8a85-4cf8-b858-01b736747bcc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.037879] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34330621-a7cf-496d-8737-c80007e31e7c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.062073] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 540.069188] env[61006]: DEBUG oslo.service.loopingcall [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.070134] env[61006]: DEBUG nova.network.neutron [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.071289] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 540.071683] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7572d918-f6e3-4fdc-830b-06ed0922d846 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.089938] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.094852] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 540.094852] env[61006]: value = "task-1336823" [ 540.094852] env[61006]: _type = "Task" [ 540.094852] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.112108] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336823, 'name': CreateVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.127412] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.128078] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.128078] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.128078] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.128385] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.128385] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.128484] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.128784] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.128784] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.128939] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.129116] env[61006]: DEBUG nova.virt.hardware [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.130399] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c34a83-98f3-4efe-9c42-659f0062e6ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.138288] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0317d9-147f-471b-9d07-95adc76aad83 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.185926] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.247037] env[61006]: INFO nova.compute.manager [-] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Took 1.06 seconds to deallocate network for instance. [ 540.250614] env[61006]: DEBUG nova.compute.claims [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.250614] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.418949] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.470877] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.489571] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Successfully created port: 07fcb90e-16ea-4824-83b2-de3be7d3992c {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.573203] env[61006]: DEBUG oslo_concurrency.lockutils [req-18c49b0f-7dff-4ea5-957d-ae2373f327c2 req-63a188fd-5a64-4899-b6a1-f30b1f3ee289 service nova] Releasing lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.616501] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336823, 'name': CreateVM_Task, 'duration_secs': 0.277965} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.619230] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 540.619838] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.619995] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.620321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 540.620579] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03482d1f-2b7b-4ce8-b795-7e68ce015706 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.625584] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 540.625584] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d93258-ee9d-816d-34fc-5c05f5dfbe1f" [ 540.625584] env[61006]: _type = "Task" [ 540.625584] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.638517] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d93258-ee9d-816d-34fc-5c05f5dfbe1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.667554] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77aa6cd-de83-49cb-b08f-59f5728d0d01 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.676092] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae956282-7c3f-4bf9-beb9-1cbc78dc6925 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.713500] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50a8e15-2895-4290-85fd-a2e910ddc029 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.724682] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3ed131-c11b-4174-9d3e-b9d67d36899f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.729561] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.739751] env[61006]: DEBUG nova.compute.provider_tree [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.742708] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.743271] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Releasing lock "refresh_cache-3769d396-ef6a-436b-ab75-faf66e4da813" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.743550] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 540.743838] env[61006]: DEBUG nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.743838] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.768897] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.921375] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Releasing lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.921810] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.922008] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.922598] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e6e9714-5f0a-48f4-8022-5a406e11bd47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.933973] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af69667-a5c4-4f56-8e52-c9eef0ff0869 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.956823] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa649a9c-8745-4188-80b8-09d62cd882e5 could not be found. [ 540.957060] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.957253] env[61006]: INFO nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 540.957733] env[61006]: DEBUG oslo.service.loopingcall [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.957962] env[61006]: DEBUG nova.compute.manager [-] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.958070] env[61006]: DEBUG nova.network.neutron [-] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.984872] env[61006]: DEBUG nova.network.neutron [-] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.140772] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d93258-ee9d-816d-34fc-5c05f5dfbe1f, 'name': SearchDatastore_Task, 'duration_secs': 0.015631} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.141112] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.141340] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 541.141570] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.141707] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.141872] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.142299] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1cd8a59-cd3f-424f-bc58-704f2f9de3d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.153872] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.156177] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 541.156992] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-491c83e0-6124-40c2-bed1-ef01af250291 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.163463] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 541.163463] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1fccf-d313-1a3f-0b74-f93f62f3cf7f" [ 541.163463] env[61006]: _type = "Task" [ 541.163463] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.177815] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1fccf-d313-1a3f-0b74-f93f62f3cf7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.249841] env[61006]: DEBUG nova.scheduler.client.report [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.272315] env[61006]: DEBUG nova.network.neutron [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.489393] env[61006]: DEBUG nova.network.neutron [-] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.677974] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1fccf-d313-1a3f-0b74-f93f62f3cf7f, 'name': SearchDatastore_Task, 'duration_secs': 0.00834} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.678798] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3061a4c-24f5-43a7-8854-68cdbb8edafb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.684718] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 541.684718] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e7e7fd-b48e-c35c-4638-66cfebca631c" [ 541.684718] env[61006]: _type = "Task" [ 541.684718] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.694037] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e7e7fd-b48e-c35c-4638-66cfebca631c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.756942] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.757877] env[61006]: ERROR nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Traceback (most recent call last): [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.driver.spawn(context, instance, image_meta, [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] vm_ref = self.build_virtual_machine(instance, [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.757877] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] for vif in network_info: [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self._sync_wrapper(fn, *args, **kwargs) [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.wait() [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self[:] = self._gt.wait() [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self._exit_event.wait() [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] result = hub.switch() [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.758217] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return self.greenlet.switch() [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] result = function(*args, **kwargs) [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] return func(*args, **kwargs) [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise e [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] nwinfo = self.network_api.allocate_for_instance( [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] created_port_ids = self._update_ports_for_instance( [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] with excutils.save_and_reraise_exception(): [ 541.758565] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] self.force_reraise() [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise self.value [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] updated_port = self._update_port( [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] _ensure_no_port_binding_failure(port) [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] raise exception.PortBindingFailed(port_id=port['id']) [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] nova.exception.PortBindingFailed: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. [ 541.758887] env[61006]: ERROR nova.compute.manager [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] [ 541.759171] env[61006]: DEBUG nova.compute.utils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.759854] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.196s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.765018] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Build of instance 06d7a63f-e04b-49d0-8827-ef25a7c9d11d was re-scheduled: Binding failed for port 3518a742-741e-4531-9e63-a8e3e8d13f1e, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 541.765018] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 541.765018] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.765018] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquired lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.765254] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.774733] env[61006]: INFO nova.compute.manager [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] Took 1.03 seconds to deallocate network for instance. [ 541.996565] env[61006]: INFO nova.compute.manager [-] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Took 1.04 seconds to deallocate network for instance. [ 541.999134] env[61006]: DEBUG nova.compute.claims [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.999335] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.196617] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e7e7fd-b48e-c35c-4638-66cfebca631c, 'name': SearchDatastore_Task, 'duration_secs': 0.008418} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.197136] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.197400] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 542.197680] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b9f3744-4ee6-4ee2-ad51-36979a64a756 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.208287] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 542.208287] env[61006]: value = "task-1336824" [ 542.208287] env[61006]: _type = "Task" [ 542.208287] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.223196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquiring lock "69ca58c7-1dc1-43d2-9c31-986e2a881801" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.223432] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "69ca58c7-1dc1-43d2-9c31-986e2a881801" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.229434] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336824, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.326801] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.479561] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.588493] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d1180d-6f43-473c-98af-6842c1719ba8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.599081] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f697686-8ef4-4af6-b460-8d45e3317893 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.638320] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c304b94a-c248-47f1-a5fd-b67e39a15c48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.645653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30705e2d-38b3-46f9-8cb2-d48710b2117a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.664714] env[61006]: DEBUG nova.compute.provider_tree [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.716202] env[61006]: ERROR nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 542.716202] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.716202] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.716202] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.716202] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.716202] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.716202] env[61006]: ERROR nova.compute.manager raise self.value [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.716202] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.716202] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.716202] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.717582] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.717582] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.717582] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 542.717582] env[61006]: ERROR nova.compute.manager [ 542.717582] env[61006]: Traceback (most recent call last): [ 542.717582] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.717582] env[61006]: listener.cb(fileno) [ 542.717582] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.717582] env[61006]: result = function(*args, **kwargs) [ 542.717582] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.717582] env[61006]: return func(*args, **kwargs) [ 542.717582] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.717582] env[61006]: raise e [ 542.717582] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.717582] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 542.717582] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.717582] env[61006]: created_port_ids = self._update_ports_for_instance( [ 542.717582] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.717582] env[61006]: with excutils.save_and_reraise_exception(): [ 542.717582] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.717582] env[61006]: self.force_reraise() [ 542.717582] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.717582] env[61006]: raise self.value [ 542.717582] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.717582] env[61006]: updated_port = self._update_port( [ 542.717582] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.717582] env[61006]: _ensure_no_port_binding_failure(port) [ 542.717582] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.717582] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.718691] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 542.718691] env[61006]: Removing descriptor: 20 [ 542.722038] env[61006]: ERROR nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Traceback (most recent call last): [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] yield resources [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.driver.spawn(context, instance, image_meta, [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] vm_ref = self.build_virtual_machine(instance, [ 542.722038] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] for vif in network_info: [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self._sync_wrapper(fn, *args, **kwargs) [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.wait() [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self[:] = self._gt.wait() [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self._exit_event.wait() [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.722562] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] result = hub.switch() [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self.greenlet.switch() [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] result = function(*args, **kwargs) [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return func(*args, **kwargs) [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise e [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] nwinfo = self.network_api.allocate_for_instance( [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] created_port_ids = self._update_ports_for_instance( [ 542.722990] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] with excutils.save_and_reraise_exception(): [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.force_reraise() [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise self.value [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] updated_port = self._update_port( [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] _ensure_no_port_binding_failure(port) [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise exception.PortBindingFailed(port_id=port['id']) [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 542.723320] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] [ 542.723640] env[61006]: INFO nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Terminating instance [ 542.723640] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336824, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.726861] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquiring lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.726861] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquired lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.726861] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 542.754811] env[61006]: DEBUG nova.compute.manager [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Received event network-vif-deleted-7046e781-c7d6-4859-81be-ff7f55f4856b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.755036] env[61006]: DEBUG nova.compute.manager [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Received event network-changed-3dd5dc5a-191d-4c53-a527-b28e518fbf2a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.756126] env[61006]: DEBUG nova.compute.manager [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Refreshing instance network info cache due to event network-changed-3dd5dc5a-191d-4c53-a527-b28e518fbf2a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.756424] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] Acquiring lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.756579] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] Acquired lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.756737] env[61006]: DEBUG nova.network.neutron [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Refreshing network info cache for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.808412] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquiring lock "a7c45039-297a-4aa7-b24b-baaa2bf2e29f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.808657] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "a7c45039-297a-4aa7-b24b-baaa2bf2e29f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.813816] env[61006]: INFO nova.scheduler.client.report [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Deleted allocations for instance 3769d396-ef6a-436b-ab75-faf66e4da813 [ 542.986979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Releasing lock "refresh_cache-06d7a63f-e04b-49d0-8827-ef25a7c9d11d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.987250] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 542.987508] env[61006]: DEBUG nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.987582] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 543.029858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "8b80965f-b981-431b-849c-4ac243ddd3c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.029858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "8b80965f-b981-431b-849c-4ac243ddd3c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.170718] env[61006]: DEBUG nova.scheduler.client.report [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.207967] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.220726] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336824, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519676} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.221308] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 543.221518] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 543.224158] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5d05002-efa3-4e78-b418-6191a12a186c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.234317] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 543.234317] env[61006]: value = "task-1336825" [ 543.234317] env[61006]: _type = "Task" [ 543.234317] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.248279] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.263226] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.298481] env[61006]: DEBUG nova.network.neutron [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.329790] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ec57d60-b9ee-4bf2-b310-ead4bf432fde tempest-ImagesOneServerNegativeTestJSON-312865452 tempest-ImagesOneServerNegativeTestJSON-312865452-project-member] Lock "3769d396-ef6a-436b-ab75-faf66e4da813" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.479s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.330209] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "3769d396-ef6a-436b-ab75-faf66e4da813" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.300s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.330404] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 3769d396-ef6a-436b-ab75-faf66e4da813] During sync_power_state the instance has a pending task (spawning). Skip. [ 543.330572] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "3769d396-ef6a-436b-ab75-faf66e4da813" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.402606] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.683092] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.683092] env[61006]: ERROR nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Traceback (most recent call last): [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.driver.spawn(context, instance, image_meta, [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.683092] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] vm_ref = self.build_virtual_machine(instance, [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] for vif in network_info: [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self._sync_wrapper(fn, *args, **kwargs) [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.wait() [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self[:] = self._gt.wait() [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self._exit_event.wait() [ 543.684524] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] result = hub.switch() [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return self.greenlet.switch() [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] result = function(*args, **kwargs) [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] return func(*args, **kwargs) [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise e [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] nwinfo = self.network_api.allocate_for_instance( [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.685066] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] created_port_ids = self._update_ports_for_instance( [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] with excutils.save_and_reraise_exception(): [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] self.force_reraise() [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise self.value [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] updated_port = self._update_port( [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] _ensure_no_port_binding_failure(port) [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.688059] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] raise exception.PortBindingFailed(port_id=port['id']) [ 543.688935] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] nova.exception.PortBindingFailed: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. [ 543.688935] env[61006]: ERROR nova.compute.manager [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] [ 543.688935] env[61006]: DEBUG nova.compute.utils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.688935] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.135s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.688935] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.688935] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 543.689156] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.172s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.689156] env[61006]: INFO nova.compute.claims [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 543.689156] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Build of instance 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94 was re-scheduled: Binding failed for port 9759c4cb-2acb-431c-aa62-4718e447227b, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.689156] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.689156] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.689345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.689790] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.691311] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcbf149-e5ef-46f3-96a9-0edd2f1a42b5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.700973] env[61006]: DEBUG nova.network.neutron [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.702952] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68da7771-1e03-4b87-8675-752bd212cf67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.721257] env[61006]: DEBUG nova.network.neutron [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.723426] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85efc70-7fd3-460d-95bb-aadfa89d4824 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.733483] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1428e7f2-f07f-4a90-ada6-641752edf128 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.781346] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181464MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 543.781346] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.784582] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068478} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.785218] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 543.786075] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df5a86f-25a5-4baf-a8c1-43fdd801f0a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.809139] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 543.809878] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88aecfd1-7242-4f34-b39e-17ac81c1648c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.832142] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 543.832142] env[61006]: value = "task-1336826" [ 543.832142] env[61006]: _type = "Task" [ 543.832142] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.836422] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 543.846062] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336826, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.857826] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquiring lock "1afdbe38-96ab-4e38-a327-2e7f4a304a4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.858381] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "1afdbe38-96ab-4e38-a327-2e7f4a304a4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.086869] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Releasing lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.086869] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.086869] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.086869] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b83e9e5-e16c-4d79-be37-dca7d11e37a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.095961] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23fd9f59-effb-4292-8e8e-dc0f964086d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.119768] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f3a466c4-3b75-4f0c-8c17-076ccd1e258e could not be found. [ 544.120025] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.120225] env[61006]: INFO nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 544.120530] env[61006]: DEBUG oslo.service.loopingcall [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.121067] env[61006]: DEBUG nova.compute.manager [-] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.121067] env[61006]: DEBUG nova.network.neutron [-] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.144567] env[61006]: DEBUG nova.network.neutron [-] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.209212] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] Releasing lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.209212] env[61006]: DEBUG nova.compute.manager [req-1e44f0e7-6519-4c0e-ae11-d48a9a7e4d9a req-2321a74f-24b8-44d9-aeac-9af7cda12e9c service nova] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Received event network-vif-deleted-3dd5dc5a-191d-4c53-a527-b28e518fbf2a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.228054] env[61006]: INFO nova.compute.manager [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] Took 1.24 seconds to deallocate network for instance. [ 544.234664] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.347473] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336826, 'name': ReconfigVM_Task, 'duration_secs': 0.26881} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.352026] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfigured VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 544.352026] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e12f770-846b-44c1-a773-4bf561dabf2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.357648] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 544.357648] env[61006]: value = "task-1336827" [ 544.357648] env[61006]: _type = "Task" [ 544.357648] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.369684] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336827, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.379956] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.436743] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.530898] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquiring lock "70e6f255-680f-4e92-a2fe-254127e70b77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.531389] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "70e6f255-680f-4e92-a2fe-254127e70b77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.647976] env[61006]: DEBUG nova.network.neutron [-] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.867853] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336827, 'name': Rename_Task, 'duration_secs': 0.126683} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.870344] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 544.872563] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0390e4b-f59f-476a-bbe5-db090a49309b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.878289] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 544.878289] env[61006]: value = "task-1336828" [ 544.878289] env[61006]: _type = "Task" [ 544.878289] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.887351] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336828, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.939210] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.939210] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 544.939210] env[61006]: DEBUG nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.939924] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.958065] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.965115] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c637a5a-9d84-446a-a875-50b634aa0783 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.975948] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ec0334-4beb-4c38-bbc4-21feba1b829c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.004587] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fcce8b-d12d-41df-b45c-4e45e3d128eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.012640] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe6bce2-0420-4dcb-b398-050872971cbb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.027789] env[61006]: DEBUG nova.compute.provider_tree [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.151747] env[61006]: INFO nova.compute.manager [-] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Took 1.03 seconds to deallocate network for instance. [ 545.154036] env[61006]: DEBUG nova.compute.claims [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.154241] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.277246] env[61006]: INFO nova.scheduler.client.report [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Deleted allocations for instance 06d7a63f-e04b-49d0-8827-ef25a7c9d11d [ 545.396099] env[61006]: DEBUG oslo_vmware.api [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336828, 'name': PowerOnVM_Task, 'duration_secs': 0.49887} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.396244] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 545.396501] env[61006]: DEBUG nova.compute.manager [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.397920] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfd6526-d744-4ad7-9814-521c600b79d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.462623] env[61006]: DEBUG nova.network.neutron [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.505163] env[61006]: DEBUG nova.compute.manager [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Received event network-changed-07fcb90e-16ea-4824-83b2-de3be7d3992c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.505444] env[61006]: DEBUG nova.compute.manager [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Refreshing instance network info cache due to event network-changed-07fcb90e-16ea-4824-83b2-de3be7d3992c. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.505568] env[61006]: DEBUG oslo_concurrency.lockutils [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] Acquiring lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.505706] env[61006]: DEBUG oslo_concurrency.lockutils [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] Acquired lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.505860] env[61006]: DEBUG nova.network.neutron [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Refreshing network info cache for port 07fcb90e-16ea-4824-83b2-de3be7d3992c {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.534531] env[61006]: DEBUG nova.scheduler.client.report [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.610211] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "f26bffa0-31b3-4a12-aa1a-3440032bd355" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.610211] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "f26bffa0-31b3-4a12-aa1a-3440032bd355" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.791226] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8427cb5-24cb-4a5d-a35a-6fea427d0ab9 tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.376s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.791226] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.760s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.791226] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 06d7a63f-e04b-49d0-8827-ef25a7c9d11d] During sync_power_state the instance has a pending task (spawning). Skip. [ 545.791226] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "06d7a63f-e04b-49d0-8827-ef25a7c9d11d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.922766] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.967269] env[61006]: INFO nova.compute.manager [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] Took 1.03 seconds to deallocate network for instance. [ 546.039100] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.039100] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.041169] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.687s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.049901] env[61006]: DEBUG nova.network.neutron [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.292914] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.351064] env[61006]: DEBUG nova.network.neutron [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.546505] env[61006]: DEBUG nova.compute.utils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.547843] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 546.701428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "e43ab4da-0b5d-4bde-a027-da603152bd72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.701807] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "e43ab4da-0b5d-4bde-a027-da603152bd72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.815472] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.856511] env[61006]: DEBUG oslo_concurrency.lockutils [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] Releasing lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.856786] env[61006]: DEBUG nova.compute.manager [req-3a45c1d0-9579-4f91-b710-80c07402d608 req-1b98a9bb-2a72-42b9-a514-3d99d571e954 service nova] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Received event network-vif-deleted-07fcb90e-16ea-4824-83b2-de3be7d3992c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.885837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fa3048-57a2-41b1-9448-a6ff8786647f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.894555] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8e4ea2-dcbf-4287-99b2-197c4dc1e4cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.927074] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59632497-2c36-422e-8d2e-e39a20cadec5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.935359] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21614389-fb72-4069-a532-7561f6bc63a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.953558] env[61006]: DEBUG nova.compute.provider_tree [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.001470] env[61006]: INFO nova.scheduler.client.report [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted allocations for instance 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94 [ 547.058285] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.251377] env[61006]: INFO nova.compute.manager [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Rebuilding instance [ 547.317795] env[61006]: DEBUG nova.compute.manager [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 547.318964] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9901224-b49b-4ae1-8ceb-28c385111980 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.458417] env[61006]: DEBUG nova.scheduler.client.report [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.511741] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b33410c0-fee0-4512-9357-81f6b3a1df60 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.568s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.514499] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.483s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.514712] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94] During sync_power_state the instance has a pending task (spawning). Skip. [ 547.514936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "3d4ef67d-7fcf-4ffc-a75c-a133b1ab9c94" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.835560] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 547.835668] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16da2504-9984-4680-a7b8-dbba058a5ff6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.847242] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 547.847242] env[61006]: value = "task-1336829" [ 547.847242] env[61006]: _type = "Task" [ 547.847242] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.859139] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336829, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.965025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.968430] env[61006]: ERROR nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Traceback (most recent call last): [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.driver.spawn(context, instance, image_meta, [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] vm_ref = self.build_virtual_machine(instance, [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.968430] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] for vif in network_info: [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self._sync_wrapper(fn, *args, **kwargs) [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.wait() [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self[:] = self._gt.wait() [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self._exit_event.wait() [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] result = hub.switch() [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.968894] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return self.greenlet.switch() [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] result = function(*args, **kwargs) [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] return func(*args, **kwargs) [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise e [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] nwinfo = self.network_api.allocate_for_instance( [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] created_port_ids = self._update_ports_for_instance( [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] with excutils.save_and_reraise_exception(): [ 547.969551] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] self.force_reraise() [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise self.value [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] updated_port = self._update_port( [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] _ensure_no_port_binding_failure(port) [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] raise exception.PortBindingFailed(port_id=port['id']) [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] nova.exception.PortBindingFailed: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. [ 547.969999] env[61006]: ERROR nova.compute.manager [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] [ 547.970334] env[61006]: DEBUG nova.compute.utils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.970334] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.723s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.971876] env[61006]: INFO nova.compute.claims [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.976109] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Build of instance 0ace1533-286c-42e2-afde-e99c77fdd747 was re-scheduled: Binding failed for port 1181ed42-1c4d-4ea7-a4b9-437f44a01bdc, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.977115] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.977115] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.977244] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquired lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.977357] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.019325] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.071258] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.117473] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.117710] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.117859] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.118061] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.118209] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.118389] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.118595] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.118744] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.118900] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.119060] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.119223] env[61006]: DEBUG nova.virt.hardware [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.120131] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63eb7733-f25a-4cab-8b5a-f4d05a798e29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.132284] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da48aa5-268c-42be-b80f-8497c9c2db15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.148966] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.155423] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Creating folder: Project (015e9970444b4a218e905e830c5ee59a). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 548.155740] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-002b4e46-c656-4dae-a72d-0f04d9d60e56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.165632] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Created folder: Project (015e9970444b4a218e905e830c5ee59a) in parent group-v285275. [ 548.166035] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Creating folder: Instances. Parent ref: group-v285280. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 548.166313] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09640dad-f2e7-4528-a14b-0b1896c77848 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.174953] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Created folder: Instances in parent group-v285280. [ 548.174953] env[61006]: DEBUG oslo.service.loopingcall [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.174953] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 548.175139] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f7e227e-b4cc-4803-999f-cdfc808278b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.193524] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.193524] env[61006]: value = "task-1336832" [ 548.193524] env[61006]: _type = "Task" [ 548.193524] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.201757] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336832, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.356957] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336829, 'name': PowerOffVM_Task, 'duration_secs': 0.15686} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.357293] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 548.357572] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.358752] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba48b680-eb24-466e-ad29-f7ac20dbf6fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.366304] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 548.366756] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5915b88c-bd99-4fa1-a13f-b7d927c5b780 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.390471] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 548.390606] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 548.390808] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Deleting the datastore file [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 548.391098] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3825748-870f-4b36-8abd-31d9fd307e33 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.403106] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 548.403106] env[61006]: value = "task-1336834" [ 548.403106] env[61006]: _type = "Task" [ 548.403106] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.411492] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.503020] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.550132] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.575382] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.710315] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336832, 'name': CreateVM_Task, 'duration_secs': 0.281831} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.710549] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 548.711407] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.711928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.712617] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 548.712861] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ded868-e3e3-4cd4-9a11-a61645b87e08 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.719130] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 548.719130] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520e011f-a8ef-283c-0897-a0349461be11" [ 548.719130] env[61006]: _type = "Task" [ 548.719130] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.729435] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520e011f-a8ef-283c-0897-a0349461be11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.917026] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137195} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.917793] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.917793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 548.917793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.080876] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Releasing lock "refresh_cache-0ace1533-286c-42e2-afde-e99c77fdd747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.081190] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.081320] env[61006]: DEBUG nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.081486] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.101287] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.239111] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520e011f-a8ef-283c-0897-a0349461be11, 'name': SearchDatastore_Task, 'duration_secs': 0.025584} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.240232] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.240232] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 549.240445] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.240619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.240823] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 549.241137] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76cf7a59-4931-4163-8ec1-a5167f53e74c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.255761] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 549.255761] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 549.255761] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-603dd67d-3c81-4cf6-9ff6-7ceaa386e2ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.261262] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 549.261262] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d5a11-0da3-6914-4108-afc3e0c88276" [ 549.261262] env[61006]: _type = "Task" [ 549.261262] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.273195] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d5a11-0da3-6914-4108-afc3e0c88276, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.371634] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297f2447-bc61-42e4-82ea-a02b9a9cdf67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.386037] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a00a30f-85da-477a-9665-32bde376abd2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.429522] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2c6e33-f667-4793-b19c-9638a4f03dac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.442030] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9cc338-1a3d-4c97-80f1-0d7d3c3f0755 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.461509] env[61006]: DEBUG nova.compute.provider_tree [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.608721] env[61006]: DEBUG nova.network.neutron [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.772980] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d5a11-0da3-6914-4108-afc3e0c88276, 'name': SearchDatastore_Task, 'duration_secs': 0.008725} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.774064] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e0719da-58fa-4cf1-a784-2088b992f258 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.785060] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 549.785060] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fba97c-1fda-e40c-f05a-e3a9e8b865da" [ 549.785060] env[61006]: _type = "Task" [ 549.785060] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.793641] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fba97c-1fda-e40c-f05a-e3a9e8b865da, 'name': SearchDatastore_Task, 'duration_secs': 0.00871} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.794025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.794493] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 549.794607] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29b0d84c-dfee-4016-83da-e6bbf0d71a6f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.802941] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 549.802941] env[61006]: value = "task-1336835" [ 549.802941] env[61006]: _type = "Task" [ 549.802941] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.811026] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.922842] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "de11f24e-7aab-4a68-a099-c5a8ca32259b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.923148] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "de11f24e-7aab-4a68-a099-c5a8ca32259b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.966725] env[61006]: DEBUG nova.scheduler.client.report [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.974376] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.974673] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.974881] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.975268] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.975526] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.975727] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.976030] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.976232] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.976403] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.976586] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.978529] env[61006]: DEBUG nova.virt.hardware [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.978529] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34fa014-0637-424d-82e3-3c2b62f6c17b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.988392] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2460ab-6d99-4c8d-aacd-2ba4cbf8d6d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.006054] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 550.013144] env[61006]: DEBUG oslo.service.loopingcall [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.013902] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 550.014182] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-716c2ff6-9073-48ca-bbfc-b2f40c5fd2aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.032823] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 550.032823] env[61006]: value = "task-1336836" [ 550.032823] env[61006]: _type = "Task" [ 550.032823] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.041717] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336836, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.114237] env[61006]: INFO nova.compute.manager [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] Took 1.03 seconds to deallocate network for instance. [ 550.319095] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478393} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.319414] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 550.319651] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 550.319992] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39603aa2-967a-47b6-8c86-485a5ecf81e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.327088] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 550.327088] env[61006]: value = "task-1336837" [ 550.327088] env[61006]: _type = "Task" [ 550.327088] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.337950] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.483158] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.483399] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.489706] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.239s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.545997] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336836, 'name': CreateVM_Task, 'duration_secs': 0.391381} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.546219] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 550.546643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.546797] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.547134] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 550.547381] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-706ee9bf-7d4e-4b92-9af2-fcb0a45a0fa0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.552204] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 550.552204] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ccb203-33a8-43dd-0348-fe2ca2b56af9" [ 550.552204] env[61006]: _type = "Task" [ 550.552204] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.563899] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ccb203-33a8-43dd-0348-fe2ca2b56af9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.838670] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064105} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.838932] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.839970] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b97ac0-fca2-4a2c-a587-7c32fed3887e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.867401] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.867401] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a990708c-1d30-4d04-a729-34f2b5d97d0f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.893882] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 550.893882] env[61006]: value = "task-1336838" [ 550.893882] env[61006]: _type = "Task" [ 550.893882] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.904170] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336838, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.991676] env[61006]: DEBUG nova.compute.utils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.995849] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.996101] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.068702] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ccb203-33a8-43dd-0348-fe2ca2b56af9, 'name': SearchDatastore_Task, 'duration_secs': 0.00937} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.068702] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.068702] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 551.068956] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.069098] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.069411] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 551.074019] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0805243-35fe-47d7-8eb9-b52b83ca26fd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.082500] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 551.082705] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 551.084912] env[61006]: DEBUG nova.policy [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15b4d23a462344f38b0de65ad53d90c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21b5d5ecd339465bb31cc7748b88acad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.086276] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35db8d61-bbed-410f-b830-5905ace4596a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.096645] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 551.096645] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52329169-2128-adb9-b508-92812b8c6de5" [ 551.096645] env[61006]: _type = "Task" [ 551.096645] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.105774] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52329169-2128-adb9-b508-92812b8c6de5, 'name': SearchDatastore_Task, 'duration_secs': 0.008632} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.108185] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b2033af-fbb3-4b62-8061-8092f2fb4d88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.121517] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 551.121517] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52aa5190-a93d-f627-94ea-614c1d32b9e1" [ 551.121517] env[61006]: _type = "Task" [ 551.121517] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.139135] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52aa5190-a93d-f627-94ea-614c1d32b9e1, 'name': SearchDatastore_Task, 'duration_secs': 0.008844} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.140236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.140236] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 551.140538] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f46614e-dfd8-4298-87fd-f6cd1aa704cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.147499] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 551.147499] env[61006]: value = "task-1336839" [ 551.147499] env[61006]: _type = "Task" [ 551.147499] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.158031] env[61006]: INFO nova.scheduler.client.report [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Deleted allocations for instance 0ace1533-286c-42e2-afde-e99c77fdd747 [ 551.177807] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.404451] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f363d688-e90d-4e09-b768-35af8c6060a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.415186] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336838, 'name': ReconfigVM_Task, 'duration_secs': 0.271273} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.418755] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.419596] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3ca2bca-107b-4551-b5b1-a706e7832f49 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.423249] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600d7f89-4fe6-40d6-8737-be04d95c6bb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.468417] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f11173d-b213-4d7c-bcd9-a3151f7de8c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.474283] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 551.474283] env[61006]: value = "task-1336840" [ 551.474283] env[61006]: _type = "Task" [ 551.474283] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.482622] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1900ab-3e9c-4e93-9f7b-148292e92e21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.493226] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336840, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.506904] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.509498] env[61006]: DEBUG nova.compute.provider_tree [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.545871] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.546764] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.658855] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475107} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.659669] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 551.659669] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 551.659669] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efc547e9-27f2-47a7-8b10-2a40bd736256 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.665553] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 551.665553] env[61006]: value = "task-1336845" [ 551.665553] env[61006]: _type = "Task" [ 551.665553] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.673593] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b68be6a7-6c20-4c7c-b169-6f0f9b244e1e tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "0ace1533-286c-42e2-afde-e99c77fdd747" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.602s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.674035] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.675469] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "0ace1533-286c-42e2-afde-e99c77fdd747" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 19.644s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.675469] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0ace1533-286c-42e2-afde-e99c77fdd747] During sync_power_state the instance has a pending task (spawning). Skip. [ 551.677219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "0ace1533-286c-42e2-afde-e99c77fdd747" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.690738] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Successfully created port: 8679e42c-8b17-41fa-8c3f-9538291d2611 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.984698] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336840, 'name': Rename_Task, 'duration_secs': 0.25225} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.984970] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 551.985869] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3453a2fe-eecf-4092-ae61-ffe34dfef8e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.997504] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 551.997504] env[61006]: value = "task-1336846" [ 551.997504] env[61006]: _type = "Task" [ 551.997504] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.002817] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.017141] env[61006]: DEBUG nova.scheduler.client.report [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.180342] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.181299] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068771} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.181811] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 552.182663] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d75a1a-97fc-4dbc-8eb3-a5bbd5a584de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.211845] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 552.212299] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aedd4dfb-8bca-4d1e-a1cd-24ebafce9816 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.235365] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 552.235365] env[61006]: value = "task-1336847" [ 552.235365] env[61006]: _type = "Task" [ 552.235365] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.245600] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.509785] env[61006]: DEBUG oslo_vmware.api [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336846, 'name': PowerOnVM_Task, 'duration_secs': 0.424988} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.509785] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 552.509956] env[61006]: INFO nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Took 4.44 seconds to spawn the instance on the hypervisor. [ 552.510146] env[61006]: DEBUG nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.510959] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf281e1-548b-4004-a82f-eae545952167 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.524165] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.527720] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.528341] env[61006]: ERROR nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Traceback (most recent call last): [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.driver.spawn(context, instance, image_meta, [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] vm_ref = self.build_virtual_machine(instance, [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.528341] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] for vif in network_info: [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self._sync_wrapper(fn, *args, **kwargs) [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.wait() [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self[:] = self._gt.wait() [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self._exit_event.wait() [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] result = hub.switch() [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.528747] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return self.greenlet.switch() [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] result = function(*args, **kwargs) [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] return func(*args, **kwargs) [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise e [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] nwinfo = self.network_api.allocate_for_instance( [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] created_port_ids = self._update_ports_for_instance( [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] with excutils.save_and_reraise_exception(): [ 552.529142] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] self.force_reraise() [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise self.value [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] updated_port = self._update_port( [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] _ensure_no_port_binding_failure(port) [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] raise exception.PortBindingFailed(port_id=port['id']) [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] nova.exception.PortBindingFailed: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. [ 552.529476] env[61006]: ERROR nova.compute.manager [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] [ 552.529793] env[61006]: DEBUG nova.compute.utils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.534276] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.789s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.535908] env[61006]: INFO nova.compute.claims [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.539387] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Build of instance 583945d7-2f35-48b1-9bb7-26358a2b0d2f was re-scheduled: Binding failed for port 7046e781-c7d6-4859-81be-ff7f55f4856b, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.540286] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.540387] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquiring lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.540489] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Acquired lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.541144] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.572788] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.572788] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.572920] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.573118] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.573261] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.573520] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.573743] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.573893] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.574064] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.574224] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.574384] env[61006]: DEBUG nova.virt.hardware [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.576220] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8201fd6a-bdf1-4571-b943-eacb4baf9448 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.586127] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fe4cf7-d4ef-4fc8-a9d1-17393c9e5c8e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.716463] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.746432] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.800860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "1fdae918-be3c-441d-8fee-ff5a2a32373c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.801399] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "1fdae918-be3c-441d-8fee-ff5a2a32373c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.933200] env[61006]: ERROR nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 552.933200] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.933200] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.933200] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.933200] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.933200] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.933200] env[61006]: ERROR nova.compute.manager raise self.value [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.933200] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 552.933200] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.933200] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 552.933642] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.933642] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 552.933642] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 552.933642] env[61006]: ERROR nova.compute.manager [ 552.933642] env[61006]: Traceback (most recent call last): [ 552.933642] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 552.933642] env[61006]: listener.cb(fileno) [ 552.933642] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.933642] env[61006]: result = function(*args, **kwargs) [ 552.933642] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.933642] env[61006]: return func(*args, **kwargs) [ 552.933642] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.933642] env[61006]: raise e [ 552.933642] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.933642] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 552.933642] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.933642] env[61006]: created_port_ids = self._update_ports_for_instance( [ 552.933642] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.933642] env[61006]: with excutils.save_and_reraise_exception(): [ 552.933642] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.933642] env[61006]: self.force_reraise() [ 552.933642] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.933642] env[61006]: raise self.value [ 552.933642] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.933642] env[61006]: updated_port = self._update_port( [ 552.933642] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.933642] env[61006]: _ensure_no_port_binding_failure(port) [ 552.933642] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.933642] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 552.934384] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 552.934384] env[61006]: Removing descriptor: 17 [ 552.934384] env[61006]: ERROR nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Traceback (most recent call last): [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] yield resources [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.driver.spawn(context, instance, image_meta, [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.934384] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] vm_ref = self.build_virtual_machine(instance, [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] for vif in network_info: [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self._sync_wrapper(fn, *args, **kwargs) [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.wait() [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self[:] = self._gt.wait() [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self._exit_event.wait() [ 552.934702] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] result = hub.switch() [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self.greenlet.switch() [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] result = function(*args, **kwargs) [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return func(*args, **kwargs) [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise e [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] nwinfo = self.network_api.allocate_for_instance( [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.935064] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] created_port_ids = self._update_ports_for_instance( [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] with excutils.save_and_reraise_exception(): [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.force_reraise() [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise self.value [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] updated_port = self._update_port( [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] _ensure_no_port_binding_failure(port) [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.935406] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise exception.PortBindingFailed(port_id=port['id']) [ 552.935717] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 552.935717] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] [ 552.935717] env[61006]: INFO nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Terminating instance [ 552.938495] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquiring lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.938495] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquired lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.938495] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 553.031509] env[61006]: INFO nova.compute.manager [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Took 19.54 seconds to build instance. [ 553.060964] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.120041] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.257109] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336847, 'name': ReconfigVM_Task, 'duration_secs': 0.573104} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.257109] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Reconfigured VM instance instance-00000004 to attach disk [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68/a2360f64-2bed-4c0a-9f99-54b2a34f8d68.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 553.260304] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be6082fa-3b29-46dc-83f1-9f6471dcd151 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.273569] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 553.273569] env[61006]: value = "task-1336848" [ 553.273569] env[61006]: _type = "Task" [ 553.273569] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.292055] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336848, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.459341] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.505108] env[61006]: DEBUG nova.compute.manager [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Received event network-changed-8679e42c-8b17-41fa-8c3f-9538291d2611 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.505316] env[61006]: DEBUG nova.compute.manager [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Refreshing instance network info cache due to event network-changed-8679e42c-8b17-41fa-8c3f-9538291d2611. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 553.505501] env[61006]: DEBUG oslo_concurrency.lockutils [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] Acquiring lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.533505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d69e9e12-4ade-4223-9748-596bd8a7a7ab tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.915s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.600474] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.628027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Releasing lock "refresh_cache-583945d7-2f35-48b1-9bb7-26358a2b0d2f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.628027] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 553.628027] env[61006]: DEBUG nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.628027] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.655445] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.787366] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336848, 'name': Rename_Task, 'duration_secs': 0.13536} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.790405] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 553.791076] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0c4f9b1-b52a-41a3-be4a-a7ca404185d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.797892] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Waiting for the task: (returnval){ [ 553.797892] env[61006]: value = "task-1336849" [ 553.797892] env[61006]: _type = "Task" [ 553.797892] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.811251] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.817860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "e2ef686c-b517-4121-88f5-01cad6b960b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.818268] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "e2ef686c-b517-4121-88f5-01cad6b960b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.886864] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "47124964-8057-46ba-8c57-1a9f4471402c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.887155] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.970793] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41020fbe-861f-446b-a802-3aa3d1623cdd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.982859] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7ea55d-8574-4309-b84a-917e19173fe2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.015999] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a64d0e3-0674-494f-9b68-de8e9f00f5f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.023430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4d8bb4-6d86-4ae0-bdc2-e3fd359c6de2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.039491] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 554.043386] env[61006]: DEBUG nova.compute.provider_tree [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.103550] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Releasing lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.103996] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 554.104703] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 554.105060] env[61006]: DEBUG oslo_concurrency.lockutils [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] Acquired lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.105246] env[61006]: DEBUG nova.network.neutron [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Refreshing network info cache for port 8679e42c-8b17-41fa-8c3f-9538291d2611 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.106684] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28927e54-bf09-4a2d-8a3f-b936901cdabf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.120268] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd8215e-ab97-4d75-b641-c568bfd19e73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.153117] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 488fc6b9-9398-4a8b-a471-0ba7323c9f9d could not be found. [ 554.153421] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 554.153841] env[61006]: INFO nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 554.157501] env[61006]: DEBUG oslo.service.loopingcall [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.157501] env[61006]: DEBUG nova.compute.manager [-] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.157501] env[61006]: DEBUG nova.network.neutron [-] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.167188] env[61006]: DEBUG nova.network.neutron [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.185899] env[61006]: DEBUG nova.network.neutron [-] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.312247] env[61006]: DEBUG oslo_vmware.api [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Task: {'id': task-1336849, 'name': PowerOnVM_Task, 'duration_secs': 0.451195} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.312497] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 554.312790] env[61006]: DEBUG nova.compute.manager [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 554.314868] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5adf378-3b04-46e5-88a3-c88d492f1b84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.551184] env[61006]: DEBUG nova.scheduler.client.report [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.574910] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.636241] env[61006]: DEBUG nova.network.neutron [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.671368] env[61006]: INFO nova.compute.manager [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] Took 1.05 seconds to deallocate network for instance. [ 554.688198] env[61006]: DEBUG nova.network.neutron [-] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.821040] env[61006]: DEBUG nova.network.neutron [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.838509] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.061315] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.063184] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.070240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.071s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.195301] env[61006]: INFO nova.compute.manager [-] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Took 1.04 seconds to deallocate network for instance. [ 555.196585] env[61006]: DEBUG nova.compute.claims [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 555.196585] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.327554] env[61006]: DEBUG oslo_concurrency.lockutils [req-15ec82d2-a7e9-4948-adfa-e96b55678ee0 req-d01bfdeb-2789-4c04-84ac-caa8a1f462d5 service nova] Releasing lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.491992] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquiring lock "aca01468-634e-4013-8a96-c7139f007c60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.492228] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "aca01468-634e-4013-8a96-c7139f007c60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.578081] env[61006]: DEBUG nova.compute.utils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.582296] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.582463] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 555.709133] env[61006]: DEBUG nova.policy [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fc1a81a484449d9a763599618d103f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80a6a9d0d832446dbaf4cadd067aa238', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.711669] env[61006]: INFO nova.scheduler.client.report [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Deleted allocations for instance 583945d7-2f35-48b1-9bb7-26358a2b0d2f [ 555.959135] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666ea7a9-15eb-4898-89c7-58a2b4ed6c90 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.966305] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7356ce5d-4067-45ea-8386-1f2ac3bdde0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.999485] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62844b9-f5df-42ff-9657-100a885e06b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.006898] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162b747-b4d1-43dd-8b54-fe519823e90b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.022085] env[61006]: DEBUG nova.compute.provider_tree [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.084644] env[61006]: INFO nova.compute.manager [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Rebuilding instance [ 556.086937] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.150420] env[61006]: DEBUG nova.compute.manager [req-f6d62c2a-e792-4698-bc75-1f2c2684bdaa req-64ae677a-8cf7-4d75-ae88-5a865bca822a service nova] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Received event network-vif-deleted-8679e42c-8b17-41fa-8c3f-9538291d2611 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.150924] env[61006]: DEBUG nova.compute.manager [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 556.152360] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9d7711-6877-4bd4-aacc-77f010a57ffd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.224582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-420739a8-62fb-4be9-ba8e-9b6d103fcae7 tempest-ServerExternalEventsTest-1132217556 tempest-ServerExternalEventsTest-1132217556-project-member] Lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.124s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.227404] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.194s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.227404] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 583945d7-2f35-48b1-9bb7-26358a2b0d2f] During sync_power_state the instance has a pending task (spawning). Skip. [ 556.227404] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "583945d7-2f35-48b1-9bb7-26358a2b0d2f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.519116] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Successfully created port: a92b8ec5-989f-48e0-8e38-b387516bee24 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.525596] env[61006]: DEBUG nova.scheduler.client.report [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.667721] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 556.667721] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a69fd24e-2127-4e2a-861f-53a4e0ef7bb8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.674662] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 556.674662] env[61006]: value = "task-1336852" [ 556.674662] env[61006]: _type = "Task" [ 556.674662] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.685465] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336852, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.728179] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.035975] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.036610] env[61006]: ERROR nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Traceback (most recent call last): [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.driver.spawn(context, instance, image_meta, [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] vm_ref = self.build_virtual_machine(instance, [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.036610] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] for vif in network_info: [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self._sync_wrapper(fn, *args, **kwargs) [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.wait() [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self[:] = self._gt.wait() [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self._exit_event.wait() [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] result = hub.switch() [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.037093] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return self.greenlet.switch() [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] result = function(*args, **kwargs) [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] return func(*args, **kwargs) [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise e [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] nwinfo = self.network_api.allocate_for_instance( [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] created_port_ids = self._update_ports_for_instance( [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] with excutils.save_and_reraise_exception(): [ 557.037443] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] self.force_reraise() [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise self.value [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] updated_port = self._update_port( [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] _ensure_no_port_binding_failure(port) [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] raise exception.PortBindingFailed(port_id=port['id']) [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] nova.exception.PortBindingFailed: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. [ 557.037819] env[61006]: ERROR nova.compute.manager [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] [ 557.038198] env[61006]: DEBUG nova.compute.utils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.039471] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.258s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.041381] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Build of instance fa649a9c-8745-4188-80b8-09d62cd882e5 was re-scheduled: Binding failed for port 3dd5dc5a-191d-4c53-a527-b28e518fbf2a, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.041854] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.046220] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquiring lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.046463] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Acquired lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.046730] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.101119] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.131055] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:08:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='320411292',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2012850198',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.131828] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.131828] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.131828] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.131828] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.134485] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.134781] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.136213] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.136213] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.136213] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.136213] env[61006]: DEBUG nova.virt.hardware [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.136449] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d388d3-50a1-42a0-a1ba-3e642eaabcae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.148130] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99aedd1a-d196-4a39-a82b-1489ddcc6f5d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.184755] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336852, 'name': PowerOffVM_Task, 'duration_secs': 0.125493} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.185572] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 557.186102] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.187121] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1fa7dc-8c4a-43dd-83ee-ee02148f1c3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.195769] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 557.195769] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2dbd152-9030-4ce6-85c7-4f376f165553 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.226233] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 557.226465] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 557.226674] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Deleting the datastore file [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 557.226929] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9def8bd6-a50c-4c06-a513-707194f656e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.241214] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 557.241214] env[61006]: value = "task-1336854" [ 557.241214] env[61006]: _type = "Task" [ 557.241214] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.257952] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336854, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.258781] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.512411] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.512411] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.512411] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.512747] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.512747] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.515645] env[61006]: INFO nova.compute.manager [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Terminating instance [ 557.520685] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "refresh_cache-a2360f64-2bed-4c0a-9f99-54b2a34f8d68" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.520956] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquired lock "refresh_cache-a2360f64-2bed-4c0a-9f99-54b2a34f8d68" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.521091] env[61006]: DEBUG nova.network.neutron [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.598017] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.757435] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.763336] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336854, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.763336] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.763336] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 557.763336] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.067941] env[61006]: DEBUG nova.network.neutron [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.100685] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance a2360f64-2bed-4c0a-9f99-54b2a34f8d68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.102761] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquiring lock "e42de415-1d9d-4931-a5b7-97112af42fd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.103092] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "e42de415-1d9d-4931-a5b7-97112af42fd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.129566] env[61006]: ERROR nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 558.129566] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.129566] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.129566] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.129566] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.129566] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.129566] env[61006]: ERROR nova.compute.manager raise self.value [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.129566] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.129566] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.129566] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.130075] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.130075] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.130075] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 558.130075] env[61006]: ERROR nova.compute.manager [ 558.130075] env[61006]: Traceback (most recent call last): [ 558.130075] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.130075] env[61006]: listener.cb(fileno) [ 558.130075] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.130075] env[61006]: result = function(*args, **kwargs) [ 558.130075] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.130075] env[61006]: return func(*args, **kwargs) [ 558.130075] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.130075] env[61006]: raise e [ 558.130075] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.130075] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 558.130075] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.130075] env[61006]: created_port_ids = self._update_ports_for_instance( [ 558.130075] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.130075] env[61006]: with excutils.save_and_reraise_exception(): [ 558.130075] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.130075] env[61006]: self.force_reraise() [ 558.130075] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.130075] env[61006]: raise self.value [ 558.130075] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.130075] env[61006]: updated_port = self._update_port( [ 558.130075] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.130075] env[61006]: _ensure_no_port_binding_failure(port) [ 558.130075] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.130075] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.130866] env[61006]: nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 558.130866] env[61006]: Removing descriptor: 17 [ 558.130866] env[61006]: ERROR nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Traceback (most recent call last): [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] yield resources [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.driver.spawn(context, instance, image_meta, [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.130866] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] vm_ref = self.build_virtual_machine(instance, [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] for vif in network_info: [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self._sync_wrapper(fn, *args, **kwargs) [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.wait() [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self[:] = self._gt.wait() [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self._exit_event.wait() [ 558.131225] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] result = hub.switch() [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self.greenlet.switch() [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] result = function(*args, **kwargs) [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return func(*args, **kwargs) [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise e [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] nwinfo = self.network_api.allocate_for_instance( [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.131599] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] created_port_ids = self._update_ports_for_instance( [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] with excutils.save_and_reraise_exception(): [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.force_reraise() [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise self.value [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] updated_port = self._update_port( [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] _ensure_no_port_binding_failure(port) [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.132015] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise exception.PortBindingFailed(port_id=port['id']) [ 558.132343] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 558.132343] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] [ 558.132343] env[61006]: INFO nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Terminating instance [ 558.136203] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquiring lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.136609] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquired lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.136675] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.183031] env[61006]: DEBUG nova.network.neutron [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.263337] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Releasing lock "refresh_cache-fa649a9c-8745-4188-80b8-09d62cd882e5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.263883] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 558.263994] env[61006]: DEBUG nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.264642] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 558.309980] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.608195] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance fa649a9c-8745-4188-80b8-09d62cd882e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.608195] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance f3a466c4-3b75-4f0c-8c17-076ccd1e258e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.608195] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 116d50e2-d8fa-4e26-8301-d1066d627982 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.608195] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 488fc6b9-9398-4a8b-a471-0ba7323c9f9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 558.676097] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.690063] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Releasing lock "refresh_cache-a2360f64-2bed-4c0a-9f99-54b2a34f8d68" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.690063] env[61006]: DEBUG nova.compute.manager [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.690063] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.690063] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104cd60f-ed05-4158-8b21-bc4b4bc27a7a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.703487] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 558.703487] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44993813-8f5c-4a56-9ee5-0a21e4718dba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.713587] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 558.713587] env[61006]: value = "task-1336856" [ 558.713587] env[61006]: _type = "Task" [ 558.713587] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.729368] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.769425] env[61006]: DEBUG nova.compute.manager [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Received event network-changed-a92b8ec5-989f-48e0-8e38-b387516bee24 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.769425] env[61006]: DEBUG nova.compute.manager [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Refreshing instance network info cache due to event network-changed-a92b8ec5-989f-48e0-8e38-b387516bee24. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 558.769425] env[61006]: DEBUG oslo_concurrency.lockutils [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] Acquiring lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.816142] env[61006]: DEBUG nova.network.neutron [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.818648] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.820660] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.824259] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.824259] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.824259] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.824259] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.824259] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.824515] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.824515] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.824515] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.824515] env[61006]: DEBUG nova.virt.hardware [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.825129] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cc0b1b-62e6-4e12-aaa5-daceef854bc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.838877] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a47c00c-f8cc-4975-80ac-c312f8418953 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.860464] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 558.869027] env[61006]: DEBUG oslo.service.loopingcall [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.869027] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 558.869027] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4079d23a-c2c3-456e-8bd1-cf7792c56f49 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.887220] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 558.887220] env[61006]: value = "task-1336857" [ 558.887220] env[61006]: _type = "Task" [ 558.887220] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.898538] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336857, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.917763] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.116583] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 41826fe1-8f39-479a-b9fd-51399753dfb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.116583] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 25ebb89a-1054-4447-83fc-91c59e4fc80b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 559.226907] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336856, 'name': PowerOffVM_Task, 'duration_secs': 0.13864} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.226907] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 559.226907] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 559.226907] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d25f53f0-a7af-44ba-8b22-87ea51871394 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.254207] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 559.254207] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 559.254207] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleting the datastore file [datastore2] a2360f64-2bed-4c0a-9f99-54b2a34f8d68 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.254207] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7396e6b-ea90-4754-a040-2ff21d5ce066 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.265709] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for the task: (returnval){ [ 559.265709] env[61006]: value = "task-1336859" [ 559.265709] env[61006]: _type = "Task" [ 559.265709] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.281349] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336859, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.322669] env[61006]: INFO nova.compute.manager [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] Took 1.05 seconds to deallocate network for instance. [ 559.403566] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336857, 'name': CreateVM_Task, 'duration_secs': 0.325515} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.403745] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 559.404184] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.404313] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.406149] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 559.406277] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f5e07d-8bbc-489f-9790-20566bef47d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.419628] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 559.419628] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5299b2e2-5c6d-95e4-0194-017f28f00c58" [ 559.419628] env[61006]: _type = "Task" [ 559.419628] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.424024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Releasing lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.426989] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.427300] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.428969] env[61006]: DEBUG oslo_concurrency.lockutils [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] Acquired lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.428969] env[61006]: DEBUG nova.network.neutron [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Refreshing network info cache for port a92b8ec5-989f-48e0-8e38-b387516bee24 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 559.429491] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48826cce-84d8-4281-8108-1c44b4d0ec15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.440290] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5299b2e2-5c6d-95e4-0194-017f28f00c58, 'name': SearchDatastore_Task, 'duration_secs': 0.008654} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.445432] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.445432] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 559.446239] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.446239] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.446239] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 559.447822] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7726058f-4c27-44e7-9080-cd764bd0ffb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.454133] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74046627-4ec6-4fb2-9ad0-a54486936ce2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.479476] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 559.479695] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 559.480468] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b9381a-754e-49c8-8ff3-d980bf3c9ce5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.489578] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25ebb89a-1054-4447-83fc-91c59e4fc80b could not be found. [ 559.489908] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.490031] env[61006]: INFO nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Took 0.06 seconds to destroy the instance on the hypervisor. [ 559.490381] env[61006]: DEBUG oslo.service.loopingcall [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.491490] env[61006]: DEBUG nova.compute.manager [-] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.491594] env[61006]: DEBUG nova.network.neutron [-] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.493579] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 559.493579] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520d9420-d869-bdd0-7d9a-a27e2ac21bd6" [ 559.493579] env[61006]: _type = "Task" [ 559.493579] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.504285] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520d9420-d869-bdd0-7d9a-a27e2ac21bd6, 'name': SearchDatastore_Task, 'duration_secs': 0.008722} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.505053] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32505867-4e5d-4957-8b33-465111cb2501 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.512590] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 559.512590] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cf0064-0ace-71d3-d88b-07e2bcb99a94" [ 559.512590] env[61006]: _type = "Task" [ 559.512590] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.519763] env[61006]: DEBUG nova.network.neutron [-] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.525155] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cf0064-0ace-71d3-d88b-07e2bcb99a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.620100] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 075e54ee-fc17-426d-ac3a-a632699274a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.779584] env[61006]: DEBUG oslo_vmware.api [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Task: {'id': task-1336859, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098165} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.779584] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 559.779584] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 559.779584] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.779584] env[61006]: INFO nova.compute.manager [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Took 1.09 seconds to destroy the instance on the hypervisor. [ 559.779877] env[61006]: DEBUG oslo.service.loopingcall [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.779877] env[61006]: DEBUG nova.compute.manager [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.779877] env[61006]: DEBUG nova.network.neutron [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.796534] env[61006]: DEBUG nova.network.neutron [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.959304] env[61006]: DEBUG nova.network.neutron [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.029608] env[61006]: DEBUG nova.network.neutron [-] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.037482] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cf0064-0ace-71d3-d88b-07e2bcb99a94, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.037482] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.037482] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 560.037482] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64eacbdd-644f-4a14-81a1-028be7f87b0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.051548] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 560.051548] env[61006]: value = "task-1336860" [ 560.051548] env[61006]: _type = "Task" [ 560.051548] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.062376] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.067816] env[61006]: DEBUG nova.network.neutron [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.116467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "cf4e241c-a4e3-438c-8059-c7f7cb41cf45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.116467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "cf4e241c-a4e3-438c-8059-c7f7cb41cf45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.124164] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 69ca58c7-1dc1-43d2-9c31-986e2a881801 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.300468] env[61006]: DEBUG nova.network.neutron [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.368729] env[61006]: INFO nova.scheduler.client.report [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Deleted allocations for instance fa649a9c-8745-4188-80b8-09d62cd882e5 [ 560.538822] env[61006]: INFO nova.compute.manager [-] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Took 1.04 seconds to deallocate network for instance. [ 560.544542] env[61006]: DEBUG nova.compute.claims [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.544542] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.565658] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336860, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.576464] env[61006]: DEBUG oslo_concurrency.lockutils [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] Releasing lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.576464] env[61006]: DEBUG nova.compute.manager [req-7069d0ba-bcd2-4c9c-b712-00e77e165454 req-c2191322-82f3-4f82-8d56-f884dd877d9a service nova] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Received event network-vif-deleted-a92b8ec5-989f-48e0-8e38-b387516bee24 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.627592] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance a7c45039-297a-4aa7-b24b-baaa2bf2e29f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.804069] env[61006]: INFO nova.compute.manager [-] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Took 1.02 seconds to deallocate network for instance. [ 560.878557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0424fb5c-9d96-4c95-a37d-730c9f88ba3d tempest-AttachInterfacesUnderV243Test-921783019 tempest-AttachInterfacesUnderV243Test-921783019-project-member] Lock "fa649a9c-8745-4188-80b8-09d62cd882e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.358s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.880053] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "fa649a9c-8745-4188-80b8-09d62cd882e5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 28.848s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.880420] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b758354-baa2-4faf-9a64-3daa5b9c8c73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.898293] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c7e29c-f80e-41aa-8d91-016a1256f146 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.070560] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579897} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.071225] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 561.071476] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 561.071829] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0603e1b7-f125-4f0a-966b-e4b666332672 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.082896] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 561.082896] env[61006]: value = "task-1336862" [ 561.082896] env[61006]: _type = "Task" [ 561.082896] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.096710] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.130557] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 8b80965f-b981-431b-849c-4ac243ddd3c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.312145] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.389349] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.437396] env[61006]: INFO nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: fa649a9c-8745-4188-80b8-09d62cd882e5] During the sync_power process the instance has moved from host None to host cpu-1 [ 561.437636] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "fa649a9c-8745-4188-80b8-09d62cd882e5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.558s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.597385] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.271196} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.597859] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 561.598625] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be269ed-5019-4588-b73d-9c55451a3b6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.622056] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 561.622296] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a7ce25c-f0fe-4fa5-9fb4-30a98f56e031 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.647219] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1afdbe38-96ab-4e38-a327-2e7f4a304a4c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.654254] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 561.654254] env[61006]: value = "task-1336863" [ 561.654254] env[61006]: _type = "Task" [ 561.654254] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.664351] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336863, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.927529] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.147292] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 70e6f255-680f-4e92-a2fe-254127e70b77 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.165355] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336863, 'name': ReconfigVM_Task, 'duration_secs': 0.339601} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.165355] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982/116d50e2-d8fa-4e26-8301-d1066d627982.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 562.165775] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d051937b-8dcd-46bc-bd79-35bff2519c84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.173622] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 562.173622] env[61006]: value = "task-1336864" [ 562.173622] env[61006]: _type = "Task" [ 562.173622] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.183648] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336864, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.652102] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance f26bffa0-31b3-4a12-aa1a-3440032bd355 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.692932] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336864, 'name': Rename_Task, 'duration_secs': 0.143435} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.692932] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 562.692932] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f96f1b62-0733-4585-8ec2-1d2811343a00 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.700785] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 562.700785] env[61006]: value = "task-1336865" [ 562.700785] env[61006]: _type = "Task" [ 562.700785] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.711349] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.157964] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e43ab4da-0b5d-4bde-a027-da603152bd72 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.218394] env[61006]: DEBUG oslo_vmware.api [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336865, 'name': PowerOnVM_Task, 'duration_secs': 0.466607} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.220956] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 563.220956] env[61006]: DEBUG nova.compute.manager [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 563.220956] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98053af-0b45-4b00-9390-453db66d6f79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.663732] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance de11f24e-7aab-4a68-a099-c5a8ca32259b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.748618] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.167883] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6642ffcb-cb01-4e38-a27c-bf4e4c938a17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.671357] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1fdae918-be3c-441d-8fee-ff5a2a32373c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.791427] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "116d50e2-d8fa-4e26-8301-d1066d627982" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.791796] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.791885] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "116d50e2-d8fa-4e26-8301-d1066d627982-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.792056] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.795998] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.797596] env[61006]: INFO nova.compute.manager [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Terminating instance [ 564.799680] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "refresh_cache-116d50e2-d8fa-4e26-8301-d1066d627982" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.799835] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquired lock "refresh_cache-116d50e2-d8fa-4e26-8301-d1066d627982" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.799998] env[61006]: DEBUG nova.network.neutron [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.175854] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e2ef686c-b517-4121-88f5-01cad6b960b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.339765] env[61006]: DEBUG nova.network.neutron [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.572225] env[61006]: DEBUG nova.network.neutron [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.682394] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 47124964-8057-46ba-8c57-1a9f4471402c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.078027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Releasing lock "refresh_cache-116d50e2-d8fa-4e26-8301-d1066d627982" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.078306] env[61006]: DEBUG nova.compute.manager [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.078498] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.079589] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e50dbd9-d573-4cc7-a3c2-6ba600d1f0a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.089621] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 566.090169] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1921369e-2dfe-423d-a4a4-5bdd39a75d6e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.098454] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 566.098454] env[61006]: value = "task-1336868" [ 566.098454] env[61006]: _type = "Task" [ 566.098454] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.109720] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.188678] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance aca01468-634e-4013-8a96-c7139f007c60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.612268] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336868, 'name': PowerOffVM_Task, 'duration_secs': 0.13441} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.612626] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 566.612795] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 566.615495] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d4f8750-023a-4206-98bb-a180a425394d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.646672] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 566.647617] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 566.647859] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Deleting the datastore file [datastore2] 116d50e2-d8fa-4e26-8301-d1066d627982 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 566.648153] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6429ac2-8db3-469d-8910-be46fdd8c72c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.661432] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for the task: (returnval){ [ 566.661432] env[61006]: value = "task-1336870" [ 566.661432] env[61006]: _type = "Task" [ 566.661432] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.670590] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.694028] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e42de415-1d9d-4931-a5b7-97112af42fd1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.694028] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 566.694028] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 567.115201] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe270bb-0f03-4821-b7b8-9ef7bd0f5e9d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.125830] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c898c95d-df9c-4863-8ce1-c5ef7e4ffd55 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.160828] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14de6e5b-9016-498a-bb86-16ec0d92c611 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.176507] env[61006]: DEBUG oslo_vmware.api [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Task: {'id': task-1336870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11404} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.176890] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 567.178091] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 567.178091] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.178091] env[61006]: INFO nova.compute.manager [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Took 1.10 seconds to destroy the instance on the hypervisor. [ 567.178091] env[61006]: DEBUG oslo.service.loopingcall [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.178957] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c958027-47d1-45bf-a42c-627f50c209a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.183245] env[61006]: DEBUG nova.compute.manager [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.183383] env[61006]: DEBUG nova.network.neutron [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.198964] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.200893] env[61006]: DEBUG nova.network.neutron [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.708049] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.716204] env[61006]: DEBUG nova.network.neutron [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.088992] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "5e4a8c02-be44-4139-b591-e1ed4ef52261" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.088992] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "5e4a8c02-be44-4139-b591-e1ed4ef52261" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.214571] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 568.214735] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.176s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.215176] env[61006]: INFO nova.compute.manager [-] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Took 1.03 seconds to deallocate network for instance. [ 568.215352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.836s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.217229] env[61006]: INFO nova.compute.claims [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.220085] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 568.220509] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Getting list of instances from cluster (obj){ [ 568.220509] env[61006]: value = "domain-c8" [ 568.220509] env[61006]: _type = "ClusterComputeResource" [ 568.220509] env[61006]: } {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 568.225903] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dfeb9b-a05b-4d95-b67e-0c4d4da56a01 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.245227] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Got total of 0 instances {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 568.736452] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.044439] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "09646374-a3e3-4fc6-a227-950966144315" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.044914] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "09646374-a3e3-4fc6-a227-950966144315" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.216239] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "8102112b-ee4d-4c7e-9b67-9b6a9c032f56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.216637] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "8102112b-ee4d-4c7e-9b67-9b6a9c032f56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.725010] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb8ca55-14f4-497a-887e-ffe48c1e7ff1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.735772] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe6cd1f-eadb-43de-84fe-9125635624d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.776136] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab974b2-e242-4439-8647-df75362f7e48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.786097] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52f2d8d-a887-41bd-9100-9f11be1345d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.801606] env[61006]: DEBUG nova.compute.provider_tree [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.304691] env[61006]: DEBUG nova.scheduler.client.report [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.813031] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.813031] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.815516] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.660s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.015044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquiring lock "5a2ddd3d-c939-4f20-a40d-108e805d64c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.015227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "5a2ddd3d-c939-4f20-a40d-108e805d64c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.324786] env[61006]: DEBUG nova.compute.utils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.326332] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 571.791513] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dac8b2c-463f-4e6e-8e8b-8dddf9443d12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.802984] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9124239f-f90a-4941-880d-db8c20679223 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.835985] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.842785] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6509e1ce-4490-40ef-9510-7c8bd91350e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.852273] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff978121-e9a5-4e84-9e17-601c04e72d94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.868303] env[61006]: DEBUG nova.compute.provider_tree [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.154269] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquiring lock "1542f4da-16bb-47c3-b11a-a9ac332ac917" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.154540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "1542f4da-16bb-47c3-b11a-a9ac332ac917" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.374022] env[61006]: DEBUG nova.scheduler.client.report [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.855953] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.879939] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.065s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.881844] env[61006]: ERROR nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Traceback (most recent call last): [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.driver.spawn(context, instance, image_meta, [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] vm_ref = self.build_virtual_machine(instance, [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.881844] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] for vif in network_info: [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self._sync_wrapper(fn, *args, **kwargs) [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.wait() [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self[:] = self._gt.wait() [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self._exit_event.wait() [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] result = hub.switch() [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.882301] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return self.greenlet.switch() [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] result = function(*args, **kwargs) [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] return func(*args, **kwargs) [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise e [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] nwinfo = self.network_api.allocate_for_instance( [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] created_port_ids = self._update_ports_for_instance( [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] with excutils.save_and_reraise_exception(): [ 572.882699] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] self.force_reraise() [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise self.value [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] updated_port = self._update_port( [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] _ensure_no_port_binding_failure(port) [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] raise exception.PortBindingFailed(port_id=port['id']) [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] nova.exception.PortBindingFailed: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. [ 572.883073] env[61006]: ERROR nova.compute.manager [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] [ 572.883383] env[61006]: DEBUG nova.compute.utils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.883533] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.961s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.883730] env[61006]: DEBUG nova.objects.instance [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61006) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 572.893017] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.893017] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.893017] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.893297] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.893297] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.893297] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.893297] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.893297] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.893489] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.893489] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.893489] env[61006]: DEBUG nova.virt.hardware [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.893489] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Build of instance f3a466c4-3b75-4f0c-8c17-076ccd1e258e was re-scheduled: Binding failed for port 07fcb90e-16ea-4824-83b2-de3be7d3992c, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.893489] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.893758] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquiring lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.893758] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Acquired lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.893758] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.895339] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ff8b55-babf-460b-b988-b8914da7e8c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.907045] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f66518-67f1-45d1-8e3a-74d24c8df4b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.926357] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 572.933045] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Creating folder: Project (f2aa34d9e2824cd5b02fd905a43c51e2). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 572.933407] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-299f0928-f821-4b3f-a0e0-bc50a00176e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.946785] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Created folder: Project (f2aa34d9e2824cd5b02fd905a43c51e2) in parent group-v285275. [ 572.947036] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Creating folder: Instances. Parent ref: group-v285289. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 572.947664] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94e99d4c-5148-4b56-8a87-cfbfc5040e5f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.959822] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Created folder: Instances in parent group-v285289. [ 572.959940] env[61006]: DEBUG oslo.service.loopingcall [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.960067] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 572.960279] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-601c1d1f-8d6c-41f9-94e6-91e6157127dc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.979753] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 572.979753] env[61006]: value = "task-1336874" [ 572.979753] env[61006]: _type = "Task" [ 572.979753] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.989714] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336874, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.318948] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "c0b97c94-a055-4ce3-b961-1cdb6a5955a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.319224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "c0b97c94-a055-4ce3-b961-1cdb6a5955a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.441408] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.492611] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336874, 'name': CreateVM_Task, 'duration_secs': 0.30275} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.493178] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 573.493663] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.494086] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.494436] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 573.494874] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-174ce366-3d22-4650-8981-f56f943b6650 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.503311] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 573.503311] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524c30e9-72de-b4ec-1bf2-5e23cbc94077" [ 573.503311] env[61006]: _type = "Task" [ 573.503311] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.511560] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524c30e9-72de-b4ec-1bf2-5e23cbc94077, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.608226] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.908410] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6169f2e1-4264-4c5d-ad05-f9e4497887cf tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.024s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.909668] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.094s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.917058] env[61006]: INFO nova.compute.claims [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 574.015659] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524c30e9-72de-b4ec-1bf2-5e23cbc94077, 'name': SearchDatastore_Task, 'duration_secs': 0.013015} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.015960] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.016245] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 574.016426] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.016559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.016782] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 574.017057] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74876f4c-f8c4-44d7-8538-d64a5bc6ac6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.028027] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 574.028206] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 574.029037] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78c34d8a-a496-46b7-b2ad-370bc93124ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.036294] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 574.036294] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5279d0c3-e0c8-fbe5-c4b0-ab192d11e2c4" [ 574.036294] env[61006]: _type = "Task" [ 574.036294] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.047730] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5279d0c3-e0c8-fbe5-c4b0-ab192d11e2c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.111804] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Releasing lock "refresh_cache-f3a466c4-3b75-4f0c-8c17-076ccd1e258e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.111804] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.111804] env[61006]: DEBUG nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.111804] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.136532] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.532461] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "f7325d28-ed07-4dab-b440-2daf766882de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.533096] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "f7325d28-ed07-4dab-b440-2daf766882de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.548833] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5279d0c3-e0c8-fbe5-c4b0-ab192d11e2c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009397} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.550283] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b835cdbf-9e1d-42be-bcda-274576bbeda4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.560271] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 574.560271] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52814e50-1399-4b96-adf8-64319e69e6fc" [ 574.560271] env[61006]: _type = "Task" [ 574.560271] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.571508] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52814e50-1399-4b96-adf8-64319e69e6fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.638336] env[61006]: DEBUG nova.network.neutron [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.083178] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52814e50-1399-4b96-adf8-64319e69e6fc, 'name': SearchDatastore_Task, 'duration_secs': 0.010256} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.088891] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.088891] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 41826fe1-8f39-479a-b9fd-51399753dfb5/41826fe1-8f39-479a-b9fd-51399753dfb5.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 575.089177] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7df008f9-39da-4baf-9fe6-2c702dc60866 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.101280] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 575.101280] env[61006]: value = "task-1336875" [ 575.101280] env[61006]: _type = "Task" [ 575.101280] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.117061] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.142243] env[61006]: INFO nova.compute.manager [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] [instance: f3a466c4-3b75-4f0c-8c17-076ccd1e258e] Took 1.03 seconds to deallocate network for instance. [ 575.510438] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41cb2bd-65a2-4fd8-ab83-e6a3adbfcd8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.522638] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85569f63-411d-4d43-bd6f-5817fa865d3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.561097] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d352d8-6409-44ac-9168-91741576b620 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.572491] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fae3aa-096d-4a72-918c-8777fe388e96 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.587737] env[61006]: DEBUG nova.compute.provider_tree [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.613098] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500487} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.613373] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 41826fe1-8f39-479a-b9fd-51399753dfb5/41826fe1-8f39-479a-b9fd-51399753dfb5.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 575.613580] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 575.613836] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81abbbd3-23ba-4472-83ad-500ca70d0fa0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.622446] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 575.622446] env[61006]: value = "task-1336876" [ 575.622446] env[61006]: _type = "Task" [ 575.622446] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.634915] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.091475] env[61006]: DEBUG nova.scheduler.client.report [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.132736] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067203} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.132992] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 576.137018] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a1a90f-ad43-48f1-bed8-9f80a5ae49ca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.158967] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 41826fe1-8f39-479a-b9fd-51399753dfb5/41826fe1-8f39-479a-b9fd-51399753dfb5.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 576.162656] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ddc08e8-1306-4d59-93f5-4576dcc381ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.186546] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 576.186546] env[61006]: value = "task-1336877" [ 576.186546] env[61006]: _type = "Task" [ 576.186546] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.195547] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336877, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.205403] env[61006]: INFO nova.scheduler.client.report [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Deleted allocations for instance f3a466c4-3b75-4f0c-8c17-076ccd1e258e [ 576.597035] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.598279] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 576.600396] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.053s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.601993] env[61006]: INFO nova.compute.claims [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.700722] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336877, 'name': ReconfigVM_Task, 'duration_secs': 0.289477} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.701026] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 41826fe1-8f39-479a-b9fd-51399753dfb5/41826fe1-8f39-479a-b9fd-51399753dfb5.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 576.701923] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcb22468-5c47-4420-a9cf-bded5c2e8e94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.711191] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 576.711191] env[61006]: value = "task-1336878" [ 576.711191] env[61006]: _type = "Task" [ 576.711191] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.715388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-11fcd234-aa8e-4838-8149-799d7b71eebc tempest-ServerDiagnosticsNegativeTest-792700030 tempest-ServerDiagnosticsNegativeTest-792700030-project-member] Lock "f3a466c4-3b75-4f0c-8c17-076ccd1e258e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.242s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.729809] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336878, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.758962] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquiring lock "5525a859-eb09-45a6-80d6-8a466976a7bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.758962] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "5525a859-eb09-45a6-80d6-8a466976a7bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.108986] env[61006]: DEBUG nova.compute.utils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.110779] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.111112] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.189139] env[61006]: DEBUG nova.policy [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e29413ec784446849fbace7b61fc30bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39b11c61ab1846c88d07e930936e91aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.219829] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.227357] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336878, 'name': Rename_Task, 'duration_secs': 0.15479} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.227624] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 577.228046] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78194eb8-9bdc-4cb9-9e8e-655ac0017dac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.236434] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 577.236434] env[61006]: value = "task-1336879" [ 577.236434] env[61006]: _type = "Task" [ 577.236434] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.249946] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.592788] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Successfully created port: 5841720b-70e6-4742-b9d5-a76a4e21629c {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 577.614432] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 577.754176] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336879, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.762712] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.148849] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e51568-6078-40bc-bf73-d3620dbe5d71 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.159622] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1a4e28-7e6f-490f-a21e-deda6694f948 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.201038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395b6da1-d1e7-4a1a-abaa-93543bc40b9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.211534] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350b2742-64f4-4bff-8805-25df85313072 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.228698] env[61006]: DEBUG nova.compute.provider_tree [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.254034] env[61006]: DEBUG oslo_vmware.api [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336879, 'name': PowerOnVM_Task, 'duration_secs': 0.585726} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.254034] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 578.254034] env[61006]: INFO nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Took 5.40 seconds to spawn the instance on the hypervisor. [ 578.254034] env[61006]: DEBUG nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.254034] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a65f104-e7be-4173-95fe-fc4cfd41ff66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.262240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquiring lock "d6892ecb-0fb8-44da-9362-8806c429626b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.262240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "d6892ecb-0fb8-44da-9362-8806c429626b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.622431] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.650020] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.650020] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.650020] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.650285] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.650285] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.650285] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.650285] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.650418] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.650562] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.650711] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.650879] env[61006]: DEBUG nova.virt.hardware [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.652439] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63052f5c-0f61-47ba-b106-781d734371f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.662643] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964115af-89ed-4f94-ac65-72bcfb46dc04 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.731664] env[61006]: DEBUG nova.scheduler.client.report [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.777235] env[61006]: INFO nova.compute.manager [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Took 34.43 seconds to build instance. [ 579.079322] env[61006]: ERROR nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 579.079322] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.079322] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.079322] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.079322] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.079322] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.079322] env[61006]: ERROR nova.compute.manager raise self.value [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.079322] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.079322] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.079322] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.079884] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.079884] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.079884] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 579.079884] env[61006]: ERROR nova.compute.manager [ 579.080151] env[61006]: Traceback (most recent call last): [ 579.080227] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.080227] env[61006]: listener.cb(fileno) [ 579.080227] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.080227] env[61006]: result = function(*args, **kwargs) [ 579.080227] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.080227] env[61006]: return func(*args, **kwargs) [ 579.080227] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.080227] env[61006]: raise e [ 579.080227] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.080227] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 579.080227] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.080227] env[61006]: created_port_ids = self._update_ports_for_instance( [ 579.080227] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.080227] env[61006]: with excutils.save_and_reraise_exception(): [ 579.080227] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.080227] env[61006]: self.force_reraise() [ 579.080227] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.080227] env[61006]: raise self.value [ 579.080227] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.080227] env[61006]: updated_port = self._update_port( [ 579.080227] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.080227] env[61006]: _ensure_no_port_binding_failure(port) [ 579.080227] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.080227] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.080227] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 579.080227] env[61006]: Removing descriptor: 20 [ 579.082519] env[61006]: ERROR nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Traceback (most recent call last): [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] yield resources [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.driver.spawn(context, instance, image_meta, [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] vm_ref = self.build_virtual_machine(instance, [ 579.082519] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] for vif in network_info: [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self._sync_wrapper(fn, *args, **kwargs) [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.wait() [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self[:] = self._gt.wait() [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self._exit_event.wait() [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.082889] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] result = hub.switch() [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self.greenlet.switch() [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] result = function(*args, **kwargs) [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return func(*args, **kwargs) [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise e [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] nwinfo = self.network_api.allocate_for_instance( [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] created_port_ids = self._update_ports_for_instance( [ 579.083303] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] with excutils.save_and_reraise_exception(): [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.force_reraise() [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise self.value [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] updated_port = self._update_port( [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] _ensure_no_port_binding_failure(port) [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise exception.PortBindingFailed(port_id=port['id']) [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 579.083698] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] [ 579.084266] env[61006]: INFO nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Terminating instance [ 579.087503] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquiring lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.087503] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquired lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.087503] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.133383] env[61006]: DEBUG nova.compute.manager [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Received event network-changed-5841720b-70e6-4742-b9d5-a76a4e21629c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.134622] env[61006]: DEBUG nova.compute.manager [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Refreshing instance network info cache due to event network-changed-5841720b-70e6-4742-b9d5-a76a4e21629c. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.134622] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] Acquiring lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.237201] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.237988] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.246142] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.526s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.246142] env[61006]: INFO nova.compute.claims [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.278955] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0b24fc8f-4dd8-4d50-990c-821a8e112dc3 tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.359s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.516879] env[61006]: DEBUG nova.compute.manager [None req-72a4fca6-0017-49a4-a580-a3abc90be5d8 tempest-ServerDiagnosticsV248Test-1788464779 tempest-ServerDiagnosticsV248Test-1788464779-project-admin] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 579.516879] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58ec2b7-5e48-483d-86c2-012e9fa2bed7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.527586] env[61006]: INFO nova.compute.manager [None req-72a4fca6-0017-49a4-a580-a3abc90be5d8 tempest-ServerDiagnosticsV248Test-1788464779 tempest-ServerDiagnosticsV248Test-1788464779-project-admin] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Retrieving diagnostics [ 579.528445] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c02c4a-da21-43d4-bcd9-649d96a726b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.618715] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquiring lock "ea2a73fa-3ece-4c88-b117-3b9843a4f224" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.618979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "ea2a73fa-3ece-4c88-b117-3b9843a4f224" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.632281] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.740463] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.756018] env[61006]: DEBUG nova.compute.utils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.756018] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.756018] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.785112] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.829219] env[61006]: DEBUG nova.policy [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c1761cf1edc4914b4f2ee13ef4823a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b6edaabd12d428799959f60a84df2e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.243915] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Releasing lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.244396] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.244615] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.245348] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] Acquired lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.245544] env[61006]: DEBUG nova.network.neutron [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Refreshing network info cache for port 5841720b-70e6-4742-b9d5-a76a4e21629c {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.246733] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05f2f35b-34a6-4af9-a4d6-a6185ead452a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.259938] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Successfully created port: 84ee0928-a088-48eb-a1ab-6a55c60b0007 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.262486] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 580.271629] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6836e03-84ba-4c8f-bb48-e239fe6f6ec2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.306961] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 075e54ee-fc17-426d-ac3a-a632699274a5 could not be found. [ 580.307578] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.307661] env[61006]: INFO nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Took 0.06 seconds to destroy the instance on the hypervisor. [ 580.308166] env[61006]: DEBUG oslo.service.loopingcall [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.311567] env[61006]: DEBUG nova.compute.manager [-] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.311659] env[61006]: DEBUG nova.network.neutron [-] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.315300] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.355404] env[61006]: DEBUG nova.network.neutron [-] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.802271] env[61006]: DEBUG nova.network.neutron [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.802918] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f3d43f-c81a-423f-8337-dd3411be09f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.812033] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deda65e5-2fb9-48a3-a89c-75c72fefaa84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.843687] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94181801-67ab-4948-80ab-e4a1cf51125b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.852215] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb5979b-06b2-47df-8003-ed47b2fa620c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.858404] env[61006]: DEBUG nova.network.neutron [-] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.871808] env[61006]: DEBUG nova.compute.provider_tree [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.993947] env[61006]: DEBUG nova.network.neutron [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.303914] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 581.344444] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 581.344675] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 581.344838] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.345069] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 581.345217] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.345356] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 581.345555] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 581.345709] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 581.345880] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 581.346545] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 581.346545] env[61006]: DEBUG nova.virt.hardware [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 581.348293] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4b8fce-8067-491b-a44e-0b0594c49e03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.356675] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f8b568-a820-440a-9d6c-561b14e14221 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.363081] env[61006]: INFO nova.compute.manager [-] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Took 1.05 seconds to deallocate network for instance. [ 581.369198] env[61006]: DEBUG nova.compute.claims [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.369409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.382398] env[61006]: DEBUG nova.scheduler.client.report [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.394985] env[61006]: DEBUG nova.compute.manager [req-3710d58a-dbe0-43ed-90fa-cf281a27c40e req-d56ea4ed-d6a6-4578-8b0f-8ade343c900c service nova] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Received event network-vif-deleted-5841720b-70e6-4742-b9d5-a76a4e21629c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.497398] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5dd01d2-ef61-4fbd-ae49-a8c44670235e req-3869f4bc-1ad0-44d9-8e04-42bd1d2d90a9 service nova] Releasing lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.788083] env[61006]: ERROR nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 581.788083] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.788083] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.788083] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.788083] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.788083] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.788083] env[61006]: ERROR nova.compute.manager raise self.value [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.788083] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.788083] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.788083] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.788618] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.788618] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.788618] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 581.788618] env[61006]: ERROR nova.compute.manager [ 581.788618] env[61006]: Traceback (most recent call last): [ 581.788618] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.788618] env[61006]: listener.cb(fileno) [ 581.788618] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.788618] env[61006]: result = function(*args, **kwargs) [ 581.788618] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.788618] env[61006]: return func(*args, **kwargs) [ 581.788618] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.788618] env[61006]: raise e [ 581.788618] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.788618] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 581.788618] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.788618] env[61006]: created_port_ids = self._update_ports_for_instance( [ 581.788618] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.788618] env[61006]: with excutils.save_and_reraise_exception(): [ 581.788618] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.788618] env[61006]: self.force_reraise() [ 581.788618] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.788618] env[61006]: raise self.value [ 581.788618] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.788618] env[61006]: updated_port = self._update_port( [ 581.788618] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.788618] env[61006]: _ensure_no_port_binding_failure(port) [ 581.788618] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.788618] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.789720] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 581.789720] env[61006]: Removing descriptor: 20 [ 581.789720] env[61006]: ERROR nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Traceback (most recent call last): [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] yield resources [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.driver.spawn(context, instance, image_meta, [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.789720] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] vm_ref = self.build_virtual_machine(instance, [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] for vif in network_info: [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self._sync_wrapper(fn, *args, **kwargs) [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.wait() [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self[:] = self._gt.wait() [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self._exit_event.wait() [ 581.790389] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] result = hub.switch() [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self.greenlet.switch() [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] result = function(*args, **kwargs) [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return func(*args, **kwargs) [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise e [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] nwinfo = self.network_api.allocate_for_instance( [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.790908] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] created_port_ids = self._update_ports_for_instance( [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] with excutils.save_and_reraise_exception(): [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.force_reraise() [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise self.value [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] updated_port = self._update_port( [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] _ensure_no_port_binding_failure(port) [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.791462] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise exception.PortBindingFailed(port_id=port['id']) [ 581.791995] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 581.791995] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] [ 581.791995] env[61006]: INFO nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Terminating instance [ 581.792948] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquiring lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.793055] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquired lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.793211] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.887507] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.888057] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.891664] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.316s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.897097] env[61006]: INFO nova.compute.claims [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.406066] env[61006]: DEBUG nova.compute.utils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.409037] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.409037] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 582.465730] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.546955] env[61006]: DEBUG nova.policy [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '222726e403bb401a84785e95fabd7d10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c161ae874f24036ae5c2f190cd4eb7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.672390] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.916989] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.152960] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Successfully created port: 2b818349-4fcb-4620-a43e-edfd6a5290a1 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.178351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Releasing lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.178351] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.178351] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.178351] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59e26db8-4519-4710-9546-088f11ee9479 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.193526] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a210fbd4-6c64-4145-b4cb-e5e0ca7ced07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.225718] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69ca58c7-1dc1-43d2-9c31-986e2a881801 could not be found. [ 583.226175] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.226627] env[61006]: INFO nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Took 0.05 seconds to destroy the instance on the hypervisor. [ 583.227180] env[61006]: DEBUG oslo.service.loopingcall [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.231107] env[61006]: DEBUG nova.compute.manager [-] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.231107] env[61006]: DEBUG nova.network.neutron [-] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.272335] env[61006]: DEBUG nova.network.neutron [-] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.484463] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "34f17493-0d3d-4dfd-a2ac-6adc9b388f95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.484729] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "34f17493-0d3d-4dfd-a2ac-6adc9b388f95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.490392] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b12020-04cd-4fed-8a52-e9bce5b7b409 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.499319] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeba7098-846a-4b57-b39e-6356e85760e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.534222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471c5a1a-1b5d-49fa-8c7b-a3ff4a38137e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.545177] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99c1ba7-e1a6-4870-8f01-8cdd245ad59e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.563704] env[61006]: DEBUG nova.compute.provider_tree [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.621280] env[61006]: DEBUG nova.compute.manager [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Received event network-changed-84ee0928-a088-48eb-a1ab-6a55c60b0007 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.621465] env[61006]: DEBUG nova.compute.manager [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Refreshing instance network info cache due to event network-changed-84ee0928-a088-48eb-a1ab-6a55c60b0007. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.621675] env[61006]: DEBUG oslo_concurrency.lockutils [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] Acquiring lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.621894] env[61006]: DEBUG oslo_concurrency.lockutils [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] Acquired lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.621959] env[61006]: DEBUG nova.network.neutron [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Refreshing network info cache for port 84ee0928-a088-48eb-a1ab-6a55c60b0007 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.775298] env[61006]: DEBUG nova.network.neutron [-] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.936534] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.964583] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.964858] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.965020] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.965203] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.965343] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.965484] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.965680] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.965831] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.966014] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.966217] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.966394] env[61006]: DEBUG nova.virt.hardware [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.967311] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eda75ac-1766-48a7-8e77-8d9f46aac7f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.979694] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc99e1e-9ea4-4fda-a0f1-35d2696611ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.066535] env[61006]: DEBUG nova.scheduler.client.report [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.151658] env[61006]: DEBUG nova.network.neutron [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.277940] env[61006]: INFO nova.compute.manager [-] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Took 1.05 seconds to deallocate network for instance. [ 584.284465] env[61006]: DEBUG nova.compute.claims [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.287062] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.313080] env[61006]: DEBUG nova.network.neutron [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.561201] env[61006]: ERROR nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 584.561201] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.561201] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.561201] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.561201] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.561201] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.561201] env[61006]: ERROR nova.compute.manager raise self.value [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.561201] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.561201] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.561201] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.561939] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.561939] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.561939] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 584.561939] env[61006]: ERROR nova.compute.manager [ 584.561939] env[61006]: Traceback (most recent call last): [ 584.561939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.561939] env[61006]: listener.cb(fileno) [ 584.561939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.561939] env[61006]: result = function(*args, **kwargs) [ 584.561939] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.561939] env[61006]: return func(*args, **kwargs) [ 584.561939] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.561939] env[61006]: raise e [ 584.561939] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.561939] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 584.561939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.561939] env[61006]: created_port_ids = self._update_ports_for_instance( [ 584.561939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.561939] env[61006]: with excutils.save_and_reraise_exception(): [ 584.561939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.561939] env[61006]: self.force_reraise() [ 584.561939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.561939] env[61006]: raise self.value [ 584.561939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.561939] env[61006]: updated_port = self._update_port( [ 584.561939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.561939] env[61006]: _ensure_no_port_binding_failure(port) [ 584.561939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.561939] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.562877] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 584.562877] env[61006]: Removing descriptor: 17 [ 584.562877] env[61006]: ERROR nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Traceback (most recent call last): [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] yield resources [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.driver.spawn(context, instance, image_meta, [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.562877] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] vm_ref = self.build_virtual_machine(instance, [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] for vif in network_info: [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self._sync_wrapper(fn, *args, **kwargs) [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.wait() [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self[:] = self._gt.wait() [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self._exit_event.wait() [ 584.563285] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] result = hub.switch() [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self.greenlet.switch() [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] result = function(*args, **kwargs) [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return func(*args, **kwargs) [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise e [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] nwinfo = self.network_api.allocate_for_instance( [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.563723] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] created_port_ids = self._update_ports_for_instance( [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] with excutils.save_and_reraise_exception(): [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.force_reraise() [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise self.value [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] updated_port = self._update_port( [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] _ensure_no_port_binding_failure(port) [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.564171] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise exception.PortBindingFailed(port_id=port['id']) [ 584.564857] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 584.564857] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] [ 584.564857] env[61006]: INFO nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Terminating instance [ 584.566876] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquiring lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.567061] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquired lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.567224] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.574326] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.681s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.574326] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.575698] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.739s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.575866] env[61006]: DEBUG nova.objects.instance [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61006) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 584.818234] env[61006]: DEBUG oslo_concurrency.lockutils [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] Releasing lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.819207] env[61006]: DEBUG nova.compute.manager [req-579d2239-ad38-4e68-baf8-e8a5e8ad62da req-da2db697-0184-41a8-baa0-f47c8df00cad service nova] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Received event network-vif-deleted-84ee0928-a088-48eb-a1ab-6a55c60b0007 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 585.081497] env[61006]: DEBUG nova.compute.utils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.087097] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.087097] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.094936] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.170494] env[61006]: DEBUG nova.policy [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cca6e2806aa45208ae618f6a78ccc0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe9eabeec9a941e68a9eae559e24ff4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.240501] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.588915] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 585.594624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d649d93-cd1c-4c8e-a7be-011137e98c39 tempest-ServersAdmin275Test-2015902650 tempest-ServersAdmin275Test-2015902650-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.595754] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.400s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.618344] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Successfully created port: b3ce5b16-0b00-409e-9fe7-7fe7ba230edd {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.743019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Releasing lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.743438] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.743633] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.743950] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4bc6806-4c99-46b2-b2f8-29a5cbe3a19e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.754144] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b315016c-5403-45c2-9743-19bcb91edb1b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.778389] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7c45039-297a-4aa7-b24b-baaa2bf2e29f could not be found. [ 585.778632] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 585.778849] env[61006]: INFO nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 585.779196] env[61006]: DEBUG oslo.service.loopingcall [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.779432] env[61006]: DEBUG nova.compute.manager [-] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.779521] env[61006]: DEBUG nova.network.neutron [-] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.804032] env[61006]: DEBUG nova.network.neutron [-] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.182583] env[61006]: DEBUG nova.compute.manager [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Received event network-changed-2b818349-4fcb-4620-a43e-edfd6a5290a1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.183177] env[61006]: DEBUG nova.compute.manager [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Refreshing instance network info cache due to event network-changed-2b818349-4fcb-4620-a43e-edfd6a5290a1. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.183414] env[61006]: DEBUG oslo_concurrency.lockutils [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] Acquiring lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.183550] env[61006]: DEBUG oslo_concurrency.lockutils [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] Acquired lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.184113] env[61006]: DEBUG nova.network.neutron [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Refreshing network info cache for port 2b818349-4fcb-4620-a43e-edfd6a5290a1 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.308538] env[61006]: DEBUG nova.network.neutron [-] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.602437] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 586.638737] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.639065] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.640202] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.640202] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.640202] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.640202] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.641555] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.641555] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.641555] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.641555] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.641555] env[61006]: DEBUG nova.virt.hardware [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.642248] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ca5de9-c0f4-400b-b65f-cc01a8e107c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.653920] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d15ccd-4ee9-4e2c-b6a5-089cc04a9911 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.661067] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1059e00-2210-41a2-805a-d6c924039ade {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.678808] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2c5823-ef13-4f5c-8edd-07e2b9c2d5e4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.713464] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f05f7c-5c07-4bb7-ad5a-5a67b346ece9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.722901] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea265752-51eb-45f3-9b41-f98e792c4446 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.739018] env[61006]: DEBUG nova.compute.provider_tree [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.741671] env[61006]: DEBUG nova.network.neutron [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.815690] env[61006]: INFO nova.compute.manager [-] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Took 1.04 seconds to deallocate network for instance. [ 586.818170] env[61006]: DEBUG nova.compute.claims [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 586.818395] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.065439] env[61006]: DEBUG nova.network.neutron [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.207367] env[61006]: ERROR nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 587.207367] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.207367] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.207367] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.207367] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.207367] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.207367] env[61006]: ERROR nova.compute.manager raise self.value [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.207367] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.207367] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.207367] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.208940] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.208940] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.208940] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 587.208940] env[61006]: ERROR nova.compute.manager [ 587.208940] env[61006]: Traceback (most recent call last): [ 587.208940] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.208940] env[61006]: listener.cb(fileno) [ 587.208940] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.208940] env[61006]: result = function(*args, **kwargs) [ 587.208940] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.208940] env[61006]: return func(*args, **kwargs) [ 587.208940] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.208940] env[61006]: raise e [ 587.208940] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.208940] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 587.208940] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.208940] env[61006]: created_port_ids = self._update_ports_for_instance( [ 587.208940] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.208940] env[61006]: with excutils.save_and_reraise_exception(): [ 587.208940] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.208940] env[61006]: self.force_reraise() [ 587.208940] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.208940] env[61006]: raise self.value [ 587.208940] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.208940] env[61006]: updated_port = self._update_port( [ 587.208940] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.208940] env[61006]: _ensure_no_port_binding_failure(port) [ 587.208940] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.208940] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.209818] env[61006]: nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 587.209818] env[61006]: Removing descriptor: 20 [ 587.209818] env[61006]: ERROR nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Traceback (most recent call last): [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] yield resources [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.driver.spawn(context, instance, image_meta, [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.209818] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] vm_ref = self.build_virtual_machine(instance, [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] for vif in network_info: [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self._sync_wrapper(fn, *args, **kwargs) [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.wait() [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self[:] = self._gt.wait() [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self._exit_event.wait() [ 587.210216] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] result = hub.switch() [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self.greenlet.switch() [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] result = function(*args, **kwargs) [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return func(*args, **kwargs) [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise e [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] nwinfo = self.network_api.allocate_for_instance( [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.210586] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] created_port_ids = self._update_ports_for_instance( [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] with excutils.save_and_reraise_exception(): [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.force_reraise() [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise self.value [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] updated_port = self._update_port( [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] _ensure_no_port_binding_failure(port) [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.211014] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise exception.PortBindingFailed(port_id=port['id']) [ 587.211404] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 587.211404] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] [ 587.211404] env[61006]: INFO nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Terminating instance [ 587.211404] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.211404] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.211404] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.240805] env[61006]: DEBUG nova.scheduler.client.report [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.571106] env[61006]: DEBUG oslo_concurrency.lockutils [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] Releasing lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.571390] env[61006]: DEBUG nova.compute.manager [req-423266ad-a99d-4791-8718-0b96fe1d1943 req-5ef0be33-7d5e-46cb-b522-5acd870215ce service nova] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Received event network-vif-deleted-2b818349-4fcb-4620-a43e-edfd6a5290a1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.729543] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.749428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.750815] env[61006]: ERROR nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Traceback (most recent call last): [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.driver.spawn(context, instance, image_meta, [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] vm_ref = self.build_virtual_machine(instance, [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.750815] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] for vif in network_info: [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self._sync_wrapper(fn, *args, **kwargs) [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.wait() [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self[:] = self._gt.wait() [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self._exit_event.wait() [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] result = hub.switch() [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.751267] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return self.greenlet.switch() [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] result = function(*args, **kwargs) [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] return func(*args, **kwargs) [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise e [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] nwinfo = self.network_api.allocate_for_instance( [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] created_port_ids = self._update_ports_for_instance( [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] with excutils.save_and_reraise_exception(): [ 587.751648] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] self.force_reraise() [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise self.value [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] updated_port = self._update_port( [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] _ensure_no_port_binding_failure(port) [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] raise exception.PortBindingFailed(port_id=port['id']) [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] nova.exception.PortBindingFailed: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. [ 587.752022] env[61006]: ERROR nova.compute.manager [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] [ 587.752343] env[61006]: DEBUG nova.compute.utils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.752343] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.493s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.753476] env[61006]: INFO nova.compute.claims [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.756256] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Build of instance 488fc6b9-9398-4a8b-a471-0ba7323c9f9d was re-scheduled: Binding failed for port 8679e42c-8b17-41fa-8c3f-9538291d2611, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.756491] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.756738] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquiring lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.756862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Acquired lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.757027] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.855498] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.242788] env[61006]: DEBUG nova.compute.manager [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Received event network-changed-b3ce5b16-0b00-409e-9fe7-7fe7ba230edd {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.243019] env[61006]: DEBUG nova.compute.manager [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Refreshing instance network info cache due to event network-changed-b3ce5b16-0b00-409e-9fe7-7fe7ba230edd. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.243181] env[61006]: DEBUG oslo_concurrency.lockutils [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] Acquiring lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.283385] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.358305] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.358733] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 588.358974] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.359839] env[61006]: DEBUG oslo_concurrency.lockutils [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] Acquired lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.359839] env[61006]: DEBUG nova.network.neutron [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Refreshing network info cache for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.360661] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8c16931-03ca-4891-808c-f31cd5a6e46c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.372046] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dfd7a9-d8c8-47c0-87e0-b0b081153793 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.390386] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.399589] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b80965f-b981-431b-849c-4ac243ddd3c5 could not be found. [ 588.399801] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.400014] env[61006]: INFO nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 588.400268] env[61006]: DEBUG oslo.service.loopingcall [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.401026] env[61006]: DEBUG nova.compute.manager [-] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.401266] env[61006]: DEBUG nova.network.neutron [-] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.422097] env[61006]: DEBUG nova.network.neutron [-] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.883904] env[61006]: DEBUG nova.network.neutron [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.892783] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Releasing lock "refresh_cache-488fc6b9-9398-4a8b-a471-0ba7323c9f9d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.894018] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.894018] env[61006]: DEBUG nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.894018] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.925646] env[61006]: DEBUG nova.network.neutron [-] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.926498] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.002984] env[61006]: DEBUG nova.network.neutron [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.177237] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673d084b-e3cb-45b1-b895-cf053195b800 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.188729] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b2d761-d0cb-467f-b086-b19eb8510bd5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.219129] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea12f915-9b28-4fc8-8962-560b55d6570f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.226621] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0408cd7f-8da4-4c85-86c1-53424381f631 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.240855] env[61006]: DEBUG nova.compute.provider_tree [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.428246] env[61006]: INFO nova.compute.manager [-] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Took 1.03 seconds to deallocate network for instance. [ 589.431410] env[61006]: DEBUG nova.network.neutron [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.432653] env[61006]: DEBUG nova.compute.claims [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.432822] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.508169] env[61006]: DEBUG oslo_concurrency.lockutils [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] Releasing lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.508439] env[61006]: DEBUG nova.compute.manager [req-c6a58811-48d8-4e76-8404-ff59ecea94e2 req-ce6bdb79-2d3c-4756-8800-460d1470e27d service nova] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Received event network-vif-deleted-b3ce5b16-0b00-409e-9fe7-7fe7ba230edd {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.746076] env[61006]: DEBUG nova.scheduler.client.report [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.939435] env[61006]: INFO nova.compute.manager [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] [instance: 488fc6b9-9398-4a8b-a471-0ba7323c9f9d] Took 1.04 seconds to deallocate network for instance. [ 590.058910] env[61006]: DEBUG nova.compute.manager [None req-1997204d-87f4-4d54-8b71-3d2ed21a6588 tempest-ServerDiagnosticsV248Test-1788464779 tempest-ServerDiagnosticsV248Test-1788464779-project-admin] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 590.062751] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e2dda3-f72d-4e2b-9ec0-6ce3264b99ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.070629] env[61006]: INFO nova.compute.manager [None req-1997204d-87f4-4d54-8b71-3d2ed21a6588 tempest-ServerDiagnosticsV248Test-1788464779 tempest-ServerDiagnosticsV248Test-1788464779-project-admin] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Retrieving diagnostics [ 590.071419] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c2131b-e954-4d40-9dfb-76f036e37365 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.248996] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.249446] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.252565] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.708s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.758335] env[61006]: DEBUG nova.compute.utils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.762310] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.762477] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 590.843304] env[61006]: DEBUG nova.policy [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39e80f876d434fad928dded1b11a215e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '570e7ecf2a734366a774e6702974def6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.975954] env[61006]: INFO nova.scheduler.client.report [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Deleted allocations for instance 488fc6b9-9398-4a8b-a471-0ba7323c9f9d [ 591.220135] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Successfully created port: ff68215f-1200-4a8a-862f-3e51f73c1e6e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.242214] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "41826fe1-8f39-479a-b9fd-51399753dfb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.242709] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.245087] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "41826fe1-8f39-479a-b9fd-51399753dfb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.245087] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.245087] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.245844] env[61006]: INFO nova.compute.manager [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Terminating instance [ 591.248147] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "refresh_cache-41826fe1-8f39-479a-b9fd-51399753dfb5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.248543] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquired lock "refresh_cache-41826fe1-8f39-479a-b9fd-51399753dfb5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.248903] env[61006]: DEBUG nova.network.neutron [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.268022] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.274249] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e748cf89-19c1-4581-b794-f4ab29aa7296 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.283909] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bd3ca9-34d6-4e14-83f2-47f55b36c75d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.323888] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819b5804-3693-491a-9186-c401532f04c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.334370] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3325a5-4d14-4f7c-a25e-43b0320a0092 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.355356] env[61006]: DEBUG nova.compute.provider_tree [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.488813] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dd89b04f-47c1-49d1-b3de-089760d84fd9 tempest-VolumesAssistedSnapshotsTest-2088408371 tempest-VolumesAssistedSnapshotsTest-2088408371-project-member] Lock "488fc6b9-9398-4a8b-a471-0ba7323c9f9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.669s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.772641] env[61006]: DEBUG nova.network.neutron [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.853314] env[61006]: DEBUG nova.network.neutron [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.858778] env[61006]: DEBUG nova.scheduler.client.report [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.991779] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.138561] env[61006]: DEBUG nova.compute.manager [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Received event network-changed-ff68215f-1200-4a8a-862f-3e51f73c1e6e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.142021] env[61006]: DEBUG nova.compute.manager [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Refreshing instance network info cache due to event network-changed-ff68215f-1200-4a8a-862f-3e51f73c1e6e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.142021] env[61006]: DEBUG oslo_concurrency.lockutils [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] Acquiring lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.142021] env[61006]: DEBUG oslo_concurrency.lockutils [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] Acquired lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.142021] env[61006]: DEBUG nova.network.neutron [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Refreshing network info cache for port ff68215f-1200-4a8a-862f-3e51f73c1e6e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.285978] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.324660] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.324886] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.325058] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.327363] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.328299] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.328458] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.328678] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.328842] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.329079] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.329252] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.329476] env[61006]: DEBUG nova.virt.hardware [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.330355] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659e7082-5d24-42ca-9b76-62c4e08215ca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.339683] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a7ec81-48ba-4923-a9cd-005057a17734 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.356284] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Releasing lock "refresh_cache-41826fe1-8f39-479a-b9fd-51399753dfb5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.356725] env[61006]: DEBUG nova.compute.manager [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.356926] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.357712] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabdbee0-9381-4a38-a950-fc3f1618a00a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.363205] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.111s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.363823] env[61006]: ERROR nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Traceback (most recent call last): [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.driver.spawn(context, instance, image_meta, [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] vm_ref = self.build_virtual_machine(instance, [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.363823] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] for vif in network_info: [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self._sync_wrapper(fn, *args, **kwargs) [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.wait() [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self[:] = self._gt.wait() [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self._exit_event.wait() [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] result = hub.switch() [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.364227] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return self.greenlet.switch() [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] result = function(*args, **kwargs) [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] return func(*args, **kwargs) [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise e [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] nwinfo = self.network_api.allocate_for_instance( [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] created_port_ids = self._update_ports_for_instance( [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] with excutils.save_and_reraise_exception(): [ 592.364584] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] self.force_reraise() [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise self.value [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] updated_port = self._update_port( [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] _ensure_no_port_binding_failure(port) [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] raise exception.PortBindingFailed(port_id=port['id']) [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] nova.exception.PortBindingFailed: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. [ 592.364951] env[61006]: ERROR nova.compute.manager [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] [ 592.365278] env[61006]: DEBUG nova.compute.utils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.367766] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.056s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.367990] env[61006]: DEBUG nova.objects.instance [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lazy-loading 'resources' on Instance uuid a2360f64-2bed-4c0a-9f99-54b2a34f8d68 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 592.369651] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 592.369651] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Build of instance 25ebb89a-1054-4447-83fc-91c59e4fc80b was re-scheduled: Binding failed for port a92b8ec5-989f-48e0-8e38-b387516bee24, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.370120] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.370351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquiring lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.370491] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Acquired lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.370644] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.371474] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-111f8a0d-ff44-4198-8134-ceaf6353f88e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.376428] env[61006]: ERROR nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 592.376428] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.376428] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.376428] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.376428] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.376428] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.376428] env[61006]: ERROR nova.compute.manager raise self.value [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.376428] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.376428] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.376428] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.376939] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.376939] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.376939] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 592.376939] env[61006]: ERROR nova.compute.manager [ 592.376939] env[61006]: Traceback (most recent call last): [ 592.376939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.376939] env[61006]: listener.cb(fileno) [ 592.376939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.376939] env[61006]: result = function(*args, **kwargs) [ 592.376939] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.376939] env[61006]: return func(*args, **kwargs) [ 592.376939] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.376939] env[61006]: raise e [ 592.376939] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.376939] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 592.376939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.376939] env[61006]: created_port_ids = self._update_ports_for_instance( [ 592.376939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.376939] env[61006]: with excutils.save_and_reraise_exception(): [ 592.376939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.376939] env[61006]: self.force_reraise() [ 592.376939] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.376939] env[61006]: raise self.value [ 592.376939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.376939] env[61006]: updated_port = self._update_port( [ 592.376939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.376939] env[61006]: _ensure_no_port_binding_failure(port) [ 592.376939] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.376939] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.377924] env[61006]: nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 592.377924] env[61006]: Removing descriptor: 20 [ 592.377924] env[61006]: ERROR nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Traceback (most recent call last): [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] yield resources [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.driver.spawn(context, instance, image_meta, [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.377924] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] vm_ref = self.build_virtual_machine(instance, [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] for vif in network_info: [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self._sync_wrapper(fn, *args, **kwargs) [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.wait() [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self[:] = self._gt.wait() [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self._exit_event.wait() [ 592.378579] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] result = hub.switch() [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self.greenlet.switch() [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] result = function(*args, **kwargs) [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return func(*args, **kwargs) [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise e [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] nwinfo = self.network_api.allocate_for_instance( [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.379320] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] created_port_ids = self._update_ports_for_instance( [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] with excutils.save_and_reraise_exception(): [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.force_reraise() [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise self.value [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] updated_port = self._update_port( [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] _ensure_no_port_binding_failure(port) [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.379871] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise exception.PortBindingFailed(port_id=port['id']) [ 592.380382] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 592.380382] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] [ 592.380382] env[61006]: INFO nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Terminating instance [ 592.380382] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 592.380382] env[61006]: value = "task-1336880" [ 592.380382] env[61006]: _type = "Task" [ 592.380382] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.381027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquiring lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.391029] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.518184] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.661475] env[61006]: DEBUG nova.network.neutron [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.759582] env[61006]: DEBUG nova.network.neutron [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.896913] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336880, 'name': PowerOffVM_Task, 'duration_secs': 0.128908} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.897254] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 592.897426] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 592.898000] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce14e7da-5159-4d53-9543-dda5b790ff6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.905746] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.934519] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 592.935804] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 592.936606] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Deleting the datastore file [datastore2] 41826fe1-8f39-479a-b9fd-51399753dfb5 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 592.936952] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d1b8174-910e-4df6-8b5b-290bbaed9aeb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.954623] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for the task: (returnval){ [ 592.954623] env[61006]: value = "task-1336882" [ 592.954623] env[61006]: _type = "Task" [ 592.954623] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.968665] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.063104] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.264595] env[61006]: DEBUG oslo_concurrency.lockutils [req-22767a76-c602-422f-8fac-6ef101f9b640 req-ec920c4b-658f-41d0-a2ba-c31a33d75118 service nova] Releasing lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.265163] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquired lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.265342] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.428222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1e7087-cfcb-4acc-be92-482bf5fac7d8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.437188] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca19bc0-58cf-488e-8a84-d4a981bd05ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.485568] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fda7335-1524-4bcc-a284-609070614afd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.496212] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd22a743-8604-4578-a073-f6416125347b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.500792] env[61006]: DEBUG oslo_vmware.api [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Task: {'id': task-1336882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10653} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.501149] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 593.501403] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 593.501503] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.501670] env[61006]: INFO nova.compute.manager [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 593.501901] env[61006]: DEBUG oslo.service.loopingcall [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.502493] env[61006]: DEBUG nova.compute.manager [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.502570] env[61006]: DEBUG nova.network.neutron [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.512169] env[61006]: DEBUG nova.compute.provider_tree [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.519572] env[61006]: DEBUG nova.network.neutron [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.567202] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Releasing lock "refresh_cache-25ebb89a-1054-4447-83fc-91c59e4fc80b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.567452] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.567636] env[61006]: DEBUG nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.567800] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.589467] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.931294] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.015307] env[61006]: DEBUG nova.scheduler.client.report [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.021074] env[61006]: DEBUG nova.network.neutron [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.093406] env[61006]: DEBUG nova.network.neutron [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.121094] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.196129] env[61006]: DEBUG nova.compute.manager [req-7f7a1bbd-9793-4d0d-a4c6-74b32907a3cf req-2426c855-854b-4577-bb4b-f855bf9fc2cb service nova] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Received event network-vif-deleted-ff68215f-1200-4a8a-862f-3e51f73c1e6e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.525180] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.157s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.527563] env[61006]: INFO nova.compute.manager [-] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Took 1.02 seconds to deallocate network for instance. [ 594.530915] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.601s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.530915] env[61006]: INFO nova.compute.claims [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.568029] env[61006]: INFO nova.scheduler.client.report [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Deleted allocations for instance a2360f64-2bed-4c0a-9f99-54b2a34f8d68 [ 594.597717] env[61006]: INFO nova.compute.manager [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] [instance: 25ebb89a-1054-4447-83fc-91c59e4fc80b] Took 1.03 seconds to deallocate network for instance. [ 594.626669] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Releasing lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.627384] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.627917] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 594.629174] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5ff58d5-a824-4b1f-8eda-640a0e4a15f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.642893] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a9a109-d9e8-4bfe-8635-bf2c1694949b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.674037] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1afdbe38-96ab-4e38-a327-2e7f4a304a4c could not be found. [ 594.674422] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 594.674635] env[61006]: INFO nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 594.674830] env[61006]: DEBUG oslo.service.loopingcall [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.675098] env[61006]: DEBUG nova.compute.manager [-] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.675228] env[61006]: DEBUG nova.network.neutron [-] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.694329] env[61006]: DEBUG nova.network.neutron [-] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.047829] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.086205] env[61006]: DEBUG oslo_concurrency.lockutils [None req-306d2cea-9e04-45e0-ae7f-2eccdd1d9532 tempest-ServersAdmin275Test-1124389940 tempest-ServersAdmin275Test-1124389940-project-member] Lock "a2360f64-2bed-4c0a-9f99-54b2a34f8d68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.571s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.199059] env[61006]: DEBUG nova.network.neutron [-] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.641608] env[61006]: INFO nova.scheduler.client.report [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Deleted allocations for instance 25ebb89a-1054-4447-83fc-91c59e4fc80b [ 595.699565] env[61006]: INFO nova.compute.manager [-] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Took 1.02 seconds to deallocate network for instance. [ 595.706012] env[61006]: DEBUG nova.compute.claims [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.706254] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.071180] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dcc72f-51a0-4f50-b16d-48f434f1bdb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.080104] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a719ecf-971c-4857-9364-a7954fc756f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.115554] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59009e6e-72f2-4ce1-ad12-429e5faae29f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.124583] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51a1707-1392-4b27-ba40-c00be6d359be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.140496] env[61006]: DEBUG nova.compute.provider_tree [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.152597] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c044c352-6b3c-48c7-96b6-b62a4be12cf9 tempest-ServersWithSpecificFlavorTestJSON-1944364092 tempest-ServersWithSpecificFlavorTestJSON-1944364092-project-member] Lock "25ebb89a-1054-4447-83fc-91c59e4fc80b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.253s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.644672] env[61006]: DEBUG nova.scheduler.client.report [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.655566] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.154037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.154432] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.156908] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.409s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.157112] env[61006]: DEBUG nova.objects.instance [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61006) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 597.198485] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.661461] env[61006]: DEBUG nova.compute.utils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.663222] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.663422] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.811136] env[61006]: DEBUG nova.policy [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bad27b9dec794ec5bf1d5aa605690620', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62b4cd3d83884ec3b7a9899eeab5d55f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.167328] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.172665] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ead95d50-d851-4bb4-9743-a85040b0e855 tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.173773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.438s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.174778] env[61006]: DEBUG nova.objects.instance [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lazy-loading 'resources' on Instance uuid 116d50e2-d8fa-4e26-8301-d1066d627982 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 598.227710] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Successfully created port: 8b3dbd85-04c1-47d4-a5dd-79f202ecd225 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.096086] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621db15c-5e31-45ae-ba1f-5a85e4fcd361 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.104273] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e01054b-bb1f-4b2a-8b09-4fc3b2a02233 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.143621] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb4069a-6c21-4df1-80ae-6f23b6c0e141 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.152496] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edff08d8-356e-48d6-8583-69cec3966e1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.166163] env[61006]: DEBUG nova.compute.provider_tree [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.176403] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.208978] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.208978] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.208978] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.209146] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.209146] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.209636] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.209985] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.210561] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.210857] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.211136] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.211410] env[61006]: DEBUG nova.virt.hardware [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.213011] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cef1ad-2ccf-4813-a276-daf64cce9de8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.222454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4c4448-f781-45a0-836b-2cd41ce7a4b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.605108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquiring lock "2217133d-52fb-45c4-bb84-63dde0546747" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.605108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "2217133d-52fb-45c4-bb84-63dde0546747" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.643862] env[61006]: DEBUG nova.compute.manager [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Received event network-changed-8b3dbd85-04c1-47d4-a5dd-79f202ecd225 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.644093] env[61006]: DEBUG nova.compute.manager [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Refreshing instance network info cache due to event network-changed-8b3dbd85-04c1-47d4-a5dd-79f202ecd225. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.644312] env[61006]: DEBUG oslo_concurrency.lockutils [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] Acquiring lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.644452] env[61006]: DEBUG oslo_concurrency.lockutils [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] Acquired lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.644602] env[61006]: DEBUG nova.network.neutron [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Refreshing network info cache for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 599.672588] env[61006]: DEBUG nova.scheduler.client.report [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.767512] env[61006]: ERROR nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 599.767512] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.767512] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.767512] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.767512] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.767512] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.767512] env[61006]: ERROR nova.compute.manager raise self.value [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.767512] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.767512] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.767512] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.768074] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.768074] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.768074] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 599.768074] env[61006]: ERROR nova.compute.manager [ 599.768074] env[61006]: Traceback (most recent call last): [ 599.768074] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.768074] env[61006]: listener.cb(fileno) [ 599.768074] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.768074] env[61006]: result = function(*args, **kwargs) [ 599.768074] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.768074] env[61006]: return func(*args, **kwargs) [ 599.768074] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.768074] env[61006]: raise e [ 599.768074] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.768074] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 599.768074] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.768074] env[61006]: created_port_ids = self._update_ports_for_instance( [ 599.768074] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.768074] env[61006]: with excutils.save_and_reraise_exception(): [ 599.768074] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.768074] env[61006]: self.force_reraise() [ 599.768074] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.768074] env[61006]: raise self.value [ 599.768074] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.768074] env[61006]: updated_port = self._update_port( [ 599.768074] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.768074] env[61006]: _ensure_no_port_binding_failure(port) [ 599.768074] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.768074] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.768817] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 599.768817] env[61006]: Removing descriptor: 20 [ 599.768817] env[61006]: ERROR nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Traceback (most recent call last): [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] yield resources [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.driver.spawn(context, instance, image_meta, [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.768817] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] vm_ref = self.build_virtual_machine(instance, [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] for vif in network_info: [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self._sync_wrapper(fn, *args, **kwargs) [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.wait() [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self[:] = self._gt.wait() [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self._exit_event.wait() [ 599.769076] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] result = hub.switch() [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self.greenlet.switch() [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] result = function(*args, **kwargs) [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return func(*args, **kwargs) [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise e [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] nwinfo = self.network_api.allocate_for_instance( [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.769345] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] created_port_ids = self._update_ports_for_instance( [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] with excutils.save_and_reraise_exception(): [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.force_reraise() [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise self.value [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] updated_port = self._update_port( [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] _ensure_no_port_binding_failure(port) [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.769635] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise exception.PortBindingFailed(port_id=port['id']) [ 599.769869] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 599.769869] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] [ 599.769869] env[61006]: INFO nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Terminating instance [ 599.770769] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquiring lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.180242] env[61006]: DEBUG nova.network.neutron [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.180242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.006s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.184660] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.422s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.190414] env[61006]: INFO nova.compute.claims [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.214378] env[61006]: INFO nova.scheduler.client.report [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Deleted allocations for instance 116d50e2-d8fa-4e26-8301-d1066d627982 [ 600.363328] env[61006]: DEBUG nova.network.neutron [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.734146] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55fe1ca3-038c-44a0-bd76-3a0795d462ba tempest-ServerShowV257Test-462824701 tempest-ServerShowV257Test-462824701-project-member] Lock "116d50e2-d8fa-4e26-8301-d1066d627982" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.942s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.866952] env[61006]: DEBUG oslo_concurrency.lockutils [req-5b00f4a8-a882-4775-b32f-98152a8d87a3 req-06c13b4e-993b-471d-b8b6-d9b180728dc0 service nova] Releasing lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.867456] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquired lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.867650] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.401953] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.508055] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.719841] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ef374e-7b4f-4628-ae1b-fb3c741a56ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.728644] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b006dce0-993c-43ad-ad9c-2d1688438a96 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.775640] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6747dc-00b2-4b68-83db-66f2425c1791 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.783725] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aedf7b1-cadd-4391-8089-8b7dc03c40a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.801865] env[61006]: DEBUG nova.compute.provider_tree [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.870716] env[61006]: DEBUG nova.compute.manager [req-b8e81c6b-a80f-4596-a67e-c11c344922c8 req-388bf105-827f-4bed-b2ee-4ae0adfc3645 service nova] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Received event network-vif-deleted-8b3dbd85-04c1-47d4-a5dd-79f202ecd225 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.014595] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Releasing lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.018028] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.018028] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.018309] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eacefc07-0a2b-439a-81f2-326892904842 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.034168] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a83c4f-658d-40db-93ff-c2aa78805e0c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.061772] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70e6f255-680f-4e92-a2fe-254127e70b77 could not be found. [ 602.061772] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.061879] env[61006]: INFO nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Took 0.04 seconds to destroy the instance on the hypervisor. [ 602.062054] env[61006]: DEBUG oslo.service.loopingcall [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.062307] env[61006]: DEBUG nova.compute.manager [-] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.062446] env[61006]: DEBUG nova.network.neutron [-] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.096097] env[61006]: DEBUG nova.network.neutron [-] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.307893] env[61006]: DEBUG nova.scheduler.client.report [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.599250] env[61006]: DEBUG nova.network.neutron [-] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.818038] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.633s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.818127] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.823344] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.508s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.828337] env[61006]: INFO nova.compute.claims [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.104851] env[61006]: INFO nova.compute.manager [-] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Took 1.04 seconds to deallocate network for instance. [ 603.108166] env[61006]: DEBUG nova.compute.claims [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.108560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.335334] env[61006]: DEBUG nova.compute.utils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.337060] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.337060] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 603.423352] env[61006]: DEBUG nova.policy [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.789967] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Successfully created port: 53c140f8-1546-438b-9f73-8d985e0a8065 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.843865] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.168690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "0211d621-ba0e-480b-8abd-58df829a1e39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.169094] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "0211d621-ba0e-480b-8abd-58df829a1e39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.264590] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb73d155-8d82-412c-b76f-d82b4b66a1d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.272765] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af88453e-1158-452d-abad-f692c18c66da {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.311890] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e879456-9f40-4e31-8752-c9875a189bfe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.318681] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1bad5d-a25c-4180-8a7b-761b574d448c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.335230] env[61006]: DEBUG nova.compute.provider_tree [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.547674] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquiring lock "232272f0-1bf5-436e-ae24-5efa391eef57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.547854] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "232272f0-1bf5-436e-ae24-5efa391eef57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.754440] env[61006]: DEBUG nova.compute.manager [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Received event network-changed-53c140f8-1546-438b-9f73-8d985e0a8065 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.754440] env[61006]: DEBUG nova.compute.manager [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Refreshing instance network info cache due to event network-changed-53c140f8-1546-438b-9f73-8d985e0a8065. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.754440] env[61006]: DEBUG oslo_concurrency.lockutils [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] Acquiring lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.754440] env[61006]: DEBUG oslo_concurrency.lockutils [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] Acquired lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.754440] env[61006]: DEBUG nova.network.neutron [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Refreshing network info cache for port 53c140f8-1546-438b-9f73-8d985e0a8065 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.841265] env[61006]: DEBUG nova.scheduler.client.report [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.861095] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.899565] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.899813] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.899970] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.900168] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.900298] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.900440] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.900646] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.900800] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.900962] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.901141] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.901314] env[61006]: DEBUG nova.virt.hardware [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.902446] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572fbba4-e326-4bf3-9af0-d86cf6a06af6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.911343] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e43099-8d9c-46cc-9673-85c0fd101d0c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.965227] env[61006]: ERROR nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 604.965227] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.965227] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.965227] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.965227] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.965227] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.965227] env[61006]: ERROR nova.compute.manager raise self.value [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.965227] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.965227] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.965227] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.965852] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.965852] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.965852] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 604.965852] env[61006]: ERROR nova.compute.manager [ 604.965852] env[61006]: Traceback (most recent call last): [ 604.965852] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.965852] env[61006]: listener.cb(fileno) [ 604.965852] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.965852] env[61006]: result = function(*args, **kwargs) [ 604.965852] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.965852] env[61006]: return func(*args, **kwargs) [ 604.965852] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.965852] env[61006]: raise e [ 604.965852] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.965852] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 604.965852] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.965852] env[61006]: created_port_ids = self._update_ports_for_instance( [ 604.965852] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.965852] env[61006]: with excutils.save_and_reraise_exception(): [ 604.965852] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.965852] env[61006]: self.force_reraise() [ 604.965852] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.965852] env[61006]: raise self.value [ 604.965852] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.965852] env[61006]: updated_port = self._update_port( [ 604.965852] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.965852] env[61006]: _ensure_no_port_binding_failure(port) [ 604.965852] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.965852] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.966515] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 604.966515] env[61006]: Removing descriptor: 20 [ 604.966515] env[61006]: ERROR nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Traceback (most recent call last): [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] yield resources [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.driver.spawn(context, instance, image_meta, [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.966515] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] vm_ref = self.build_virtual_machine(instance, [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] for vif in network_info: [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self._sync_wrapper(fn, *args, **kwargs) [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.wait() [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self[:] = self._gt.wait() [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self._exit_event.wait() [ 604.966774] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] result = hub.switch() [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self.greenlet.switch() [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] result = function(*args, **kwargs) [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return func(*args, **kwargs) [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise e [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] nwinfo = self.network_api.allocate_for_instance( [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.967204] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] created_port_ids = self._update_ports_for_instance( [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] with excutils.save_and_reraise_exception(): [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.force_reraise() [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise self.value [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] updated_port = self._update_port( [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] _ensure_no_port_binding_failure(port) [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.967492] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise exception.PortBindingFailed(port_id=port['id']) [ 604.967777] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 604.967777] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] [ 604.967777] env[61006]: INFO nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Terminating instance [ 604.968921] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.280641] env[61006]: DEBUG nova.network.neutron [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.345246] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.346246] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 605.350039] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.981s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.382522] env[61006]: DEBUG nova.network.neutron [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.855876] env[61006]: DEBUG nova.compute.utils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.857579] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.857663] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.888471] env[61006]: DEBUG oslo_concurrency.lockutils [req-eef428e0-4348-4be7-8242-b5e761932b84 req-362f10be-1f05-4583-bba5-d3df7317d56e service nova] Releasing lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.889040] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.889187] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.910322] env[61006]: DEBUG nova.policy [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.281061] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Successfully created port: 1c0e2aed-aad5-49e2-8346-dbd2321c27f4 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.296432] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0f9890-0a24-40c4-9312-f8c488b33139 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.305116] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30431ad5-ad4f-4000-8753-87aaecb9a6dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.342728] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810ad224-7d78-4e9e-8d3a-a5f74e9a582d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.350437] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a4bae2-6b47-45b6-bde9-c5b6d4285eb3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.363661] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.366964] env[61006]: DEBUG nova.compute.provider_tree [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.418806] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.575310] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.781018] env[61006]: DEBUG nova.compute.manager [req-36718a53-8715-404f-b38c-4fd978037c0d req-e4dbe233-f5cd-4858-8b05-ccca56a721d3 service nova] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Received event network-vif-deleted-53c140f8-1546-438b-9f73-8d985e0a8065 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.875900] env[61006]: DEBUG nova.scheduler.client.report [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.077780] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.078251] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.078449] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.080239] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-be823028-a0f2-4efa-abfd-41f802685737 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.091765] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf2bfd3-ebd4-40ab-83c0-06cda6746562 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.124111] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f26bffa0-31b3-4a12-aa1a-3440032bd355 could not be found. [ 607.124111] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.124111] env[61006]: INFO nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Took 0.04 seconds to destroy the instance on the hypervisor. [ 607.124111] env[61006]: DEBUG oslo.service.loopingcall [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.124111] env[61006]: DEBUG nova.compute.manager [-] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.124111] env[61006]: DEBUG nova.network.neutron [-] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.151817] env[61006]: DEBUG nova.network.neutron [-] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.363138] env[61006]: ERROR nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 607.363138] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.363138] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.363138] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.363138] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.363138] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.363138] env[61006]: ERROR nova.compute.manager raise self.value [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.363138] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.363138] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.363138] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.363584] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.363584] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.363584] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 607.363584] env[61006]: ERROR nova.compute.manager [ 607.363584] env[61006]: Traceback (most recent call last): [ 607.363584] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.363584] env[61006]: listener.cb(fileno) [ 607.363584] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.363584] env[61006]: result = function(*args, **kwargs) [ 607.363584] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.363584] env[61006]: return func(*args, **kwargs) [ 607.363584] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.363584] env[61006]: raise e [ 607.363584] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.363584] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 607.363584] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.363584] env[61006]: created_port_ids = self._update_ports_for_instance( [ 607.363584] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.363584] env[61006]: with excutils.save_and_reraise_exception(): [ 607.363584] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.363584] env[61006]: self.force_reraise() [ 607.363584] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.363584] env[61006]: raise self.value [ 607.363584] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.363584] env[61006]: updated_port = self._update_port( [ 607.363584] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.363584] env[61006]: _ensure_no_port_binding_failure(port) [ 607.363584] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.363584] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.364323] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 607.364323] env[61006]: Removing descriptor: 20 [ 607.380818] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 607.384866] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.384866] env[61006]: ERROR nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Traceback (most recent call last): [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.driver.spawn(context, instance, image_meta, [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.384866] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] vm_ref = self.build_virtual_machine(instance, [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] for vif in network_info: [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self._sync_wrapper(fn, *args, **kwargs) [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.wait() [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self[:] = self._gt.wait() [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self._exit_event.wait() [ 607.385404] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] result = hub.switch() [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return self.greenlet.switch() [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] result = function(*args, **kwargs) [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] return func(*args, **kwargs) [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise e [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] nwinfo = self.network_api.allocate_for_instance( [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.385689] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] created_port_ids = self._update_ports_for_instance( [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] with excutils.save_and_reraise_exception(): [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] self.force_reraise() [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise self.value [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] updated_port = self._update_port( [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] _ensure_no_port_binding_failure(port) [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.385967] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] raise exception.PortBindingFailed(port_id=port['id']) [ 607.386261] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] nova.exception.PortBindingFailed: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. [ 607.386261] env[61006]: ERROR nova.compute.manager [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] [ 607.386261] env[61006]: DEBUG nova.compute.utils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.387219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.101s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.392424] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Build of instance 075e54ee-fc17-426d-ac3a-a632699274a5 was re-scheduled: Binding failed for port 5841720b-70e6-4742-b9d5-a76a4e21629c, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.392865] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.393400] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquiring lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.393400] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Acquired lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.393400] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.408196] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 607.408196] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 607.408196] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.408343] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 607.408343] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.408343] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 607.408343] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 607.408343] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 607.408482] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 607.408482] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 607.408482] env[61006]: DEBUG nova.virt.hardware [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 607.408804] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728eace0-ae86-4b81-b817-e7a97f13257f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.418017] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b34a56-8730-4614-811c-f2462d409702 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.438160] env[61006]: ERROR nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Traceback (most recent call last): [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] yield resources [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.driver.spawn(context, instance, image_meta, [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] vm_ref = self.build_virtual_machine(instance, [ 607.438160] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] for vif in network_info: [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return self._sync_wrapper(fn, *args, **kwargs) [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.wait() [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self[:] = self._gt.wait() [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return self._exit_event.wait() [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 607.438523] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] current.throw(*self._exc) [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] result = function(*args, **kwargs) [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return func(*args, **kwargs) [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise e [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] nwinfo = self.network_api.allocate_for_instance( [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] created_port_ids = self._update_ports_for_instance( [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] with excutils.save_and_reraise_exception(): [ 607.438827] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.force_reraise() [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise self.value [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] updated_port = self._update_port( [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] _ensure_no_port_binding_failure(port) [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise exception.PortBindingFailed(port_id=port['id']) [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 607.439158] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] [ 607.439158] env[61006]: INFO nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Terminating instance [ 607.439447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.439447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.439447] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.655967] env[61006]: DEBUG nova.network.neutron [-] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.921422] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.982619] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.103764] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.121224] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.158238] env[61006]: INFO nova.compute.manager [-] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Took 1.03 seconds to deallocate network for instance. [ 608.160454] env[61006]: DEBUG nova.compute.claims [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.160631] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.328538] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994028a3-b3ad-45fa-939b-2d35df386f1e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.335871] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c971540f-2122-49b9-8e43-c0b46110ec52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.366421] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a37bde-1558-4c3c-8ab2-2f0f18b1cf56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.374432] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea14558-f163-46e8-ad23-0e5cd3971460 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.388287] env[61006]: DEBUG nova.compute.provider_tree [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.611725] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Releasing lock "refresh_cache-075e54ee-fc17-426d-ac3a-a632699274a5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.611725] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.611725] env[61006]: DEBUG nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.611725] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.624865] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.627070] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.627832] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.628167] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.628837] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0570e65-912f-4dbd-8cd2-e2e79eb62c14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.640724] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3b4dac-f81b-46bf-af25-5fed904c0815 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.664991] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e43ab4da-0b5d-4bde-a027-da603152bd72 could not be found. [ 608.668035] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.668035] env[61006]: INFO nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.668035] env[61006]: DEBUG oslo.service.loopingcall [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.668035] env[61006]: DEBUG nova.compute.manager [-] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.668035] env[61006]: DEBUG nova.network.neutron [-] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.685937] env[61006]: DEBUG nova.network.neutron [-] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.836257] env[61006]: DEBUG nova.compute.manager [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Received event network-changed-1c0e2aed-aad5-49e2-8346-dbd2321c27f4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.836446] env[61006]: DEBUG nova.compute.manager [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Refreshing instance network info cache due to event network-changed-1c0e2aed-aad5-49e2-8346-dbd2321c27f4. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.836653] env[61006]: DEBUG oslo_concurrency.lockutils [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] Acquiring lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.836790] env[61006]: DEBUG oslo_concurrency.lockutils [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] Acquired lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.836984] env[61006]: DEBUG nova.network.neutron [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Refreshing network info cache for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 608.892037] env[61006]: DEBUG nova.scheduler.client.report [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.129709] env[61006]: DEBUG nova.network.neutron [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.187047] env[61006]: DEBUG nova.network.neutron [-] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.359328] env[61006]: DEBUG nova.network.neutron [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.399028] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.399028] env[61006]: ERROR nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Traceback (most recent call last): [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.driver.spawn(context, instance, image_meta, [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.399028] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] vm_ref = self.build_virtual_machine(instance, [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] for vif in network_info: [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self._sync_wrapper(fn, *args, **kwargs) [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.wait() [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self[:] = self._gt.wait() [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self._exit_event.wait() [ 609.399420] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] result = hub.switch() [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return self.greenlet.switch() [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] result = function(*args, **kwargs) [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] return func(*args, **kwargs) [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise e [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] nwinfo = self.network_api.allocate_for_instance( [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.399699] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] created_port_ids = self._update_ports_for_instance( [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] with excutils.save_and_reraise_exception(): [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] self.force_reraise() [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise self.value [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] updated_port = self._update_port( [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] _ensure_no_port_binding_failure(port) [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.400023] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] raise exception.PortBindingFailed(port_id=port['id']) [ 609.400283] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] nova.exception.PortBindingFailed: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. [ 609.400283] env[61006]: ERROR nova.compute.manager [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] [ 609.400283] env[61006]: DEBUG nova.compute.utils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.402037] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Build of instance 69ca58c7-1dc1-43d2-9c31-986e2a881801 was re-scheduled: Binding failed for port 84ee0928-a088-48eb-a1ab-6a55c60b0007, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.402037] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.402037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquiring lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.402245] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Acquired lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.402459] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.404023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.585s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.453069] env[61006]: DEBUG nova.network.neutron [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.632027] env[61006]: INFO nova.compute.manager [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] [instance: 075e54ee-fc17-426d-ac3a-a632699274a5] Took 1.02 seconds to deallocate network for instance. [ 609.690815] env[61006]: INFO nova.compute.manager [-] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Took 1.02 seconds to deallocate network for instance. [ 609.696608] env[61006]: DEBUG nova.compute.claims [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.696813] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.928751] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.958659] env[61006]: DEBUG oslo_concurrency.lockutils [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] Releasing lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.958659] env[61006]: DEBUG nova.compute.manager [req-20135ea7-27bc-4cdd-9de1-81801a08f5be req-a8f6d3d8-c048-4c72-a2d6-12845739e0e5 service nova] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Received event network-vif-deleted-1c0e2aed-aad5-49e2-8346-dbd2321c27f4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.050541] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.363877] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4619971-e6df-41c9-9a65-1d4d36239049 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.373124] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c33c09-1d64-4324-bf0a-e4e3739d737b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.406336] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a276bb-8741-4033-9f7b-fec4d6837fbd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.414622] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f1224b-fc23-4dbc-9aaf-92ae9305d5b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.430363] env[61006]: DEBUG nova.compute.provider_tree [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.554579] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Releasing lock "refresh_cache-69ca58c7-1dc1-43d2-9c31-986e2a881801" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.554822] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.554999] env[61006]: DEBUG nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.555311] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.576809] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.675407] env[61006]: INFO nova.scheduler.client.report [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Deleted allocations for instance 075e54ee-fc17-426d-ac3a-a632699274a5 [ 610.934089] env[61006]: DEBUG nova.scheduler.client.report [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.079486] env[61006]: DEBUG nova.network.neutron [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.190359] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6adc413d-b458-498d-94ee-7634dc15e36c tempest-ServerMetadataTestJSON-443136396 tempest-ServerMetadataTestJSON-443136396-project-member] Lock "075e54ee-fc17-426d-ac3a-a632699274a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.555s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.440629] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.442030] env[61006]: ERROR nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Traceback (most recent call last): [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.driver.spawn(context, instance, image_meta, [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] vm_ref = self.build_virtual_machine(instance, [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.442030] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] for vif in network_info: [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self._sync_wrapper(fn, *args, **kwargs) [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.wait() [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self[:] = self._gt.wait() [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self._exit_event.wait() [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] result = hub.switch() [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.442357] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return self.greenlet.switch() [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] result = function(*args, **kwargs) [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] return func(*args, **kwargs) [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise e [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] nwinfo = self.network_api.allocate_for_instance( [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] created_port_ids = self._update_ports_for_instance( [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] with excutils.save_and_reraise_exception(): [ 611.442645] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] self.force_reraise() [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise self.value [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] updated_port = self._update_port( [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] _ensure_no_port_binding_failure(port) [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] raise exception.PortBindingFailed(port_id=port['id']) [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] nova.exception.PortBindingFailed: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. [ 611.442943] env[61006]: ERROR nova.compute.manager [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] [ 611.443242] env[61006]: DEBUG nova.compute.utils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.443341] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.010s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.448064] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Build of instance a7c45039-297a-4aa7-b24b-baaa2bf2e29f was re-scheduled: Binding failed for port 2b818349-4fcb-4620-a43e-edfd6a5290a1, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.448064] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.448064] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquiring lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.448064] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Acquired lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.448314] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.540669] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquiring lock "6c156407-4ac6-4e34-812d-d719f1cc33d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.540897] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "6c156407-4ac6-4e34-812d-d719f1cc33d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.585402] env[61006]: INFO nova.compute.manager [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] [instance: 69ca58c7-1dc1-43d2-9c31-986e2a881801] Took 1.03 seconds to deallocate network for instance. [ 611.691836] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.969210] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.083036] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.211023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.324837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe18953-4b7e-4b9d-bcb4-03f337eb8c7b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.332747] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5e2caf-6b10-4651-9bc1-46533bbc0209 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.362846] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3761b89-8f11-433d-a3d7-08e9cf667072 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.370465] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb37d95b-78cb-4e58-80c2-afb547393f5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.383395] env[61006]: DEBUG nova.compute.provider_tree [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.585336] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Releasing lock "refresh_cache-a7c45039-297a-4aa7-b24b-baaa2bf2e29f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.585627] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.585745] env[61006]: DEBUG nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.585912] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.601922] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.619553] env[61006]: INFO nova.scheduler.client.report [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Deleted allocations for instance 69ca58c7-1dc1-43d2-9c31-986e2a881801 [ 612.886419] env[61006]: DEBUG nova.scheduler.client.report [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.105752] env[61006]: DEBUG nova.network.neutron [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.128693] env[61006]: DEBUG oslo_concurrency.lockutils [None req-096f0387-afb3-496b-beef-7a24fceb4a85 tempest-ServersTestJSON-1776824248 tempest-ServersTestJSON-1776824248-project-member] Lock "69ca58c7-1dc1-43d2-9c31-986e2a881801" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.905s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.391940] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.948s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.392567] env[61006]: ERROR nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Traceback (most recent call last): [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.driver.spawn(context, instance, image_meta, [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] vm_ref = self.build_virtual_machine(instance, [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.392567] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] for vif in network_info: [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self._sync_wrapper(fn, *args, **kwargs) [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.wait() [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self[:] = self._gt.wait() [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self._exit_event.wait() [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] result = hub.switch() [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.392854] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return self.greenlet.switch() [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] result = function(*args, **kwargs) [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] return func(*args, **kwargs) [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise e [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] nwinfo = self.network_api.allocate_for_instance( [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] created_port_ids = self._update_ports_for_instance( [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] with excutils.save_and_reraise_exception(): [ 613.393173] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] self.force_reraise() [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise self.value [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] updated_port = self._update_port( [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] _ensure_no_port_binding_failure(port) [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] raise exception.PortBindingFailed(port_id=port['id']) [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] nova.exception.PortBindingFailed: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. [ 613.393466] env[61006]: ERROR nova.compute.manager [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] [ 613.393711] env[61006]: DEBUG nova.compute.utils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.397803] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.880s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.401567] env[61006]: INFO nova.compute.claims [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.402351] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Build of instance 8b80965f-b981-431b-849c-4ac243ddd3c5 was re-scheduled: Binding failed for port b3ce5b16-0b00-409e-9fe7-7fe7ba230edd, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.402839] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.403152] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.403240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.403391] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.608448] env[61006]: INFO nova.compute.manager [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] [instance: a7c45039-297a-4aa7-b24b-baaa2bf2e29f] Took 1.02 seconds to deallocate network for instance. [ 613.631371] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.929769] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.106636] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.176416] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.613326] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-8b80965f-b981-431b-849c-4ac243ddd3c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.613609] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.613803] env[61006]: DEBUG nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.613867] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.632238] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.238938] env[61006]: INFO nova.scheduler.client.report [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Deleted allocations for instance a7c45039-297a-4aa7-b24b-baaa2bf2e29f [ 615.244312] env[61006]: DEBUG nova.network.neutron [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.447158] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f918d824-a69c-4ec0-b1bb-f3cb7a196229 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.454833] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024934e6-dc01-4fe4-8195-df227597ecd4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.483497] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27e25a9-3496-4426-8f62-4de5219aaa4c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.493925] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5651460d-6439-4dc6-89ab-ee074fa7491b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.507144] env[61006]: DEBUG nova.compute.provider_tree [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.676485] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f120e63e-bd67-4579-a960-d4683612ed9d tempest-ServersListShow296Test-1888338655 tempest-ServersListShow296Test-1888338655-project-member] Acquiring lock "c1b4dfa5-fd20-4064-85ee-da85518a5a20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.676748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f120e63e-bd67-4579-a960-d4683612ed9d tempest-ServersListShow296Test-1888338655 tempest-ServersListShow296Test-1888338655-project-member] Lock "c1b4dfa5-fd20-4064-85ee-da85518a5a20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.747022] env[61006]: INFO nova.compute.manager [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 8b80965f-b981-431b-849c-4ac243ddd3c5] Took 1.13 seconds to deallocate network for instance. [ 615.750954] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5785e7d3-ee10-453b-aaf7-0fb264adf9c5 tempest-InstanceActionsNegativeTestJSON-488810328 tempest-InstanceActionsNegativeTestJSON-488810328-project-member] Lock "a7c45039-297a-4aa7-b24b-baaa2bf2e29f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.942s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.010490] env[61006]: DEBUG nova.scheduler.client.report [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.260061] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.519280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.122s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.519896] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.522484] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.476s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.522693] env[61006]: DEBUG nova.objects.instance [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lazy-loading 'resources' on Instance uuid 41826fe1-8f39-479a-b9fd-51399753dfb5 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.796360] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.920051] env[61006]: INFO nova.scheduler.client.report [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocations for instance 8b80965f-b981-431b-849c-4ac243ddd3c5 [ 617.029364] env[61006]: DEBUG nova.compute.utils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.031042] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.031219] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.206835] env[61006]: DEBUG nova.policy [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2a55b84c0934089b626f47127f9fbac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0032cec34cea4502862dcf3539be253b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.430955] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf1c27c2-120f-4d4e-8ae3-d81e69744793 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "8b80965f-b981-431b-849c-4ac243ddd3c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.401s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.542117] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.593949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421f4991-91e3-4092-841b-320bf1aec868 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.603032] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff383ee-a762-409e-8b09-6068fd845b6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.638708] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de751478-ac3a-4c67-9fd4-36b21bbb7a07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.646763] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de96fc58-d456-47fc-a2dc-3275128343c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.661383] env[61006]: DEBUG nova.compute.provider_tree [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.935440] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.104217] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Successfully created port: bebca07e-1a36-4b2c-ad30-ba4a3d623b58 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.164861] env[61006]: DEBUG nova.scheduler.client.report [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.499024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.558358] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.581796] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.581796] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.581796] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.581950] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.581950] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.581950] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.581950] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.589646] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.589646] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.589646] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.589646] env[61006]: DEBUG nova.virt.hardware [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.589646] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a5c1cf-005e-47a0-901e-e18ffc141265 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.601238] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8495de41-80b0-4ef0-8c2f-de774119bd2f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.669796] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.147s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.672262] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.966s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.721971] env[61006]: INFO nova.scheduler.client.report [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Deleted allocations for instance 41826fe1-8f39-479a-b9fd-51399753dfb5 [ 619.234874] env[61006]: DEBUG oslo_concurrency.lockutils [None req-294b031f-a150-4bc6-9e8e-0841cb78695b tempest-ServerDiagnosticsV248Test-1843482778 tempest-ServerDiagnosticsV248Test-1843482778-project-member] Lock "41826fe1-8f39-479a-b9fd-51399753dfb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.992s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.665943] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a209ed-5b68-491f-8fbd-6dd97d45fc57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.685888] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5e3738-9ab5-4dc1-aec3-e34e6650d43b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.717978] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4086f2d-6eb4-493d-84f9-f31d8f0acf17 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.728687] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca66c67-db3d-430c-a37a-f2ed8c0f1f81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.739165] env[61006]: DEBUG nova.compute.provider_tree [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.005426] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "fd415afa-1e5f-42ea-b288-4432f15a5503" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.005862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "fd415afa-1e5f-42ea-b288-4432f15a5503" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.059633] env[61006]: DEBUG nova.compute.manager [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Received event network-changed-bebca07e-1a36-4b2c-ad30-ba4a3d623b58 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.059816] env[61006]: DEBUG nova.compute.manager [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Refreshing instance network info cache due to event network-changed-bebca07e-1a36-4b2c-ad30-ba4a3d623b58. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 620.059982] env[61006]: DEBUG oslo_concurrency.lockutils [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] Acquiring lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.060322] env[61006]: DEBUG oslo_concurrency.lockutils [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] Acquired lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.060322] env[61006]: DEBUG nova.network.neutron [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Refreshing network info cache for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.197163] env[61006]: ERROR nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 620.197163] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.197163] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.197163] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.197163] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.197163] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.197163] env[61006]: ERROR nova.compute.manager raise self.value [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.197163] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.197163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.197163] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.198508] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.198508] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.198508] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 620.198508] env[61006]: ERROR nova.compute.manager [ 620.198508] env[61006]: Traceback (most recent call last): [ 620.198508] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.198508] env[61006]: listener.cb(fileno) [ 620.198508] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.198508] env[61006]: result = function(*args, **kwargs) [ 620.198508] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.198508] env[61006]: return func(*args, **kwargs) [ 620.198508] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.198508] env[61006]: raise e [ 620.198508] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.198508] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 620.198508] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.198508] env[61006]: created_port_ids = self._update_ports_for_instance( [ 620.198508] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.198508] env[61006]: with excutils.save_and_reraise_exception(): [ 620.198508] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.198508] env[61006]: self.force_reraise() [ 620.198508] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.198508] env[61006]: raise self.value [ 620.198508] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.198508] env[61006]: updated_port = self._update_port( [ 620.198508] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.198508] env[61006]: _ensure_no_port_binding_failure(port) [ 620.198508] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.198508] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.200077] env[61006]: nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 620.200077] env[61006]: Removing descriptor: 17 [ 620.200077] env[61006]: ERROR nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Traceback (most recent call last): [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] yield resources [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.driver.spawn(context, instance, image_meta, [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.200077] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] vm_ref = self.build_virtual_machine(instance, [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] for vif in network_info: [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self._sync_wrapper(fn, *args, **kwargs) [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.wait() [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self[:] = self._gt.wait() [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self._exit_event.wait() [ 620.200697] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] result = hub.switch() [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self.greenlet.switch() [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] result = function(*args, **kwargs) [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return func(*args, **kwargs) [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise e [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] nwinfo = self.network_api.allocate_for_instance( [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.201318] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] created_port_ids = self._update_ports_for_instance( [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] with excutils.save_and_reraise_exception(): [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.force_reraise() [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise self.value [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] updated_port = self._update_port( [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] _ensure_no_port_binding_failure(port) [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.201753] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise exception.PortBindingFailed(port_id=port['id']) [ 620.202163] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 620.202163] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] [ 620.202163] env[61006]: INFO nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Terminating instance [ 620.202163] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.243900] env[61006]: DEBUG nova.scheduler.client.report [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.586732] env[61006]: DEBUG nova.network.neutron [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.753021] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.078s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.753021] env[61006]: ERROR nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Traceback (most recent call last): [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.driver.spawn(context, instance, image_meta, [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.753021] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] vm_ref = self.build_virtual_machine(instance, [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] for vif in network_info: [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self._sync_wrapper(fn, *args, **kwargs) [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.wait() [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self[:] = self._gt.wait() [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self._exit_event.wait() [ 620.753302] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] result = hub.switch() [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return self.greenlet.switch() [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] result = function(*args, **kwargs) [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] return func(*args, **kwargs) [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise e [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] nwinfo = self.network_api.allocate_for_instance( [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.753644] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] created_port_ids = self._update_ports_for_instance( [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] with excutils.save_and_reraise_exception(): [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] self.force_reraise() [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise self.value [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] updated_port = self._update_port( [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] _ensure_no_port_binding_failure(port) [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.754087] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] raise exception.PortBindingFailed(port_id=port['id']) [ 620.754515] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] nova.exception.PortBindingFailed: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. [ 620.754515] env[61006]: ERROR nova.compute.manager [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] [ 620.754515] env[61006]: DEBUG nova.compute.utils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.754515] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.554s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.758583] env[61006]: INFO nova.compute.claims [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.761698] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Build of instance 1afdbe38-96ab-4e38-a327-2e7f4a304a4c was re-scheduled: Binding failed for port ff68215f-1200-4a8a-862f-3e51f73c1e6e, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.762193] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.762418] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquiring lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.762561] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Acquired lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.762712] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.841713] env[61006]: DEBUG nova.network.neutron [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.308374] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.345785] env[61006]: DEBUG oslo_concurrency.lockutils [req-76b33fa1-abcf-480e-b03a-874451ed75e5 req-d41622d4-f8d9-4204-9e2c-0a80a12a7737 service nova] Releasing lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.346279] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquired lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.346487] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.527213] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.716112] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 621.716357] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 621.877938] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.015096] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.031539] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Releasing lock "refresh_cache-1afdbe38-96ab-4e38-a327-2e7f4a304a4c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.031539] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.031539] env[61006]: DEBUG nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.031821] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.064869] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.096310] env[61006]: DEBUG nova.compute.manager [req-d3aac372-6a1c-434d-841c-694d45453d90 req-97cb2dc0-9296-4b04-b6f2-b75c6ab42ffc service nova] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Received event network-vif-deleted-bebca07e-1a36-4b2c-ad30-ba4a3d623b58 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.226176] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.226176] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 622.226176] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 622.239861] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eadc048-8595-4ee1-86c3-14823b3b4ac5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.248747] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c120f24a-6f17-4f91-a00c-05f3402abff3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.282039] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de91cea-bfd2-489c-81e9-c6a3077f29be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.289353] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4883cff-6342-4514-85b0-4944ef775b52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.303321] env[61006]: DEBUG nova.compute.provider_tree [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.517633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Releasing lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.518168] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.518360] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.518662] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6b84087-f723-4ec1-aad1-d722b5040ce7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.529178] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0df5f6-c1a0-4ec8-a88e-18c1499727ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.551472] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de11f24e-7aab-4a68-a099-c5a8ca32259b could not be found. [ 622.551730] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.551882] env[61006]: INFO nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 622.552151] env[61006]: DEBUG oslo.service.loopingcall [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.552361] env[61006]: DEBUG nova.compute.manager [-] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.552457] env[61006]: DEBUG nova.network.neutron [-] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.574101] env[61006]: DEBUG nova.network.neutron [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.584068] env[61006]: DEBUG nova.network.neutron [-] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.733076] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 622.733254] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 622.733388] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 622.733513] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 622.733636] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 622.733761] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Didn't find any instances for network info cache update. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 622.733963] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734141] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734287] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734435] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734580] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734760] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.734888] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 622.735639] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.806869] env[61006]: DEBUG nova.scheduler.client.report [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.076757] env[61006]: INFO nova.compute.manager [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] [instance: 1afdbe38-96ab-4e38-a327-2e7f4a304a4c] Took 1.04 seconds to deallocate network for instance. [ 623.088367] env[61006]: DEBUG nova.network.neutron [-] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.240059] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.313493] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.314132] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.317797] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.209s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.592288] env[61006]: INFO nova.compute.manager [-] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Took 1.04 seconds to deallocate network for instance. [ 623.598334] env[61006]: DEBUG nova.compute.claims [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.598334] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.752456] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "4a38192e-4fb8-4767-b1d9-3d8b59a76925" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.752821] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "4a38192e-4fb8-4767-b1d9-3d8b59a76925" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.822242] env[61006]: DEBUG nova.compute.utils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.823418] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 624.078567] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "b8fb9290-755f-40b8-af3d-c17407d0c846" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.079384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "b8fb9290-755f-40b8-af3d-c17407d0c846" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.111736] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.111965] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.115598] env[61006]: INFO nova.scheduler.client.report [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Deleted allocations for instance 1afdbe38-96ab-4e38-a327-2e7f4a304a4c [ 624.155554] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "0ad1f43f-9e98-43b7-8076-20824ccc18ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.155737] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "0ad1f43f-9e98-43b7-8076-20824ccc18ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.315814] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aba7241-02a2-4486-8b0d-62af590c66c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.325293] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff2e777-4825-4dc9-98e4-a8caa1a658a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.360445] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.364483] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40bd683-09a6-4e0f-8061-b9a191dbec64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.372216] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4672223-663e-478b-98a9-8eadc37f4787 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.389625] env[61006]: DEBUG nova.compute.provider_tree [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.444093] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "e4a5a57e-f078-48ec-afe5-204c75139c15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.444391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "e4a5a57e-f078-48ec-afe5-204c75139c15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.624103] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d139aa1d-5a89-40e3-9e93-7a2b08b81227 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572 tempest-FloatingIPsAssociationNegativeTestJSON-1687399572-project-member] Lock "1afdbe38-96ab-4e38-a327-2e7f4a304a4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.766s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.892512] env[61006]: DEBUG nova.scheduler.client.report [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.036552] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d07742fb-259a-4173-84df-c7c40838ba2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.036837] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d07742fb-259a-4173-84df-c7c40838ba2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.127109] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.230156] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "1ebc21ae-d46f-486b-952a-61324aaf353f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.230396] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "1ebc21ae-d46f-486b-952a-61324aaf353f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.374473] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.400806] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.084s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.401500] env[61006]: ERROR nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Traceback (most recent call last): [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.driver.spawn(context, instance, image_meta, [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] vm_ref = self.build_virtual_machine(instance, [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.401500] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] for vif in network_info: [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self._sync_wrapper(fn, *args, **kwargs) [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.wait() [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self[:] = self._gt.wait() [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self._exit_event.wait() [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] result = hub.switch() [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.402350] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return self.greenlet.switch() [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] result = function(*args, **kwargs) [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] return func(*args, **kwargs) [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise e [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] nwinfo = self.network_api.allocate_for_instance( [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] created_port_ids = self._update_ports_for_instance( [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] with excutils.save_and_reraise_exception(): [ 625.402683] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] self.force_reraise() [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise self.value [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] updated_port = self._update_port( [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] _ensure_no_port_binding_failure(port) [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] raise exception.PortBindingFailed(port_id=port['id']) [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] nova.exception.PortBindingFailed: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. [ 625.403657] env[61006]: ERROR nova.compute.manager [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] [ 625.403993] env[61006]: DEBUG nova.compute.utils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.406444] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.246s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.410037] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Build of instance 70e6f255-680f-4e92-a2fe-254127e70b77 was re-scheduled: Binding failed for port 8b3dbd85-04c1-47d4-a5dd-79f202ecd225, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.410224] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.410958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquiring lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.411176] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Acquired lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.411375] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.416742] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.417156] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.417468] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.417931] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.418266] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.418481] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.419089] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.419089] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.419089] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.419269] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.419415] env[61006]: DEBUG nova.virt.hardware [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.420371] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28042d5-d69b-47c4-aabc-ffbe13c0e3e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.431056] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78092732-fe6d-40cb-8d34-34735b396394 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.447478] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 625.454225] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Creating folder: Project (a4122e84e0144b7a8e07f98c44872690). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 625.454529] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-334aa3a1-90d3-49cc-85df-f6505b78b9ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.466397] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Created folder: Project (a4122e84e0144b7a8e07f98c44872690) in parent group-v285275. [ 625.466571] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Creating folder: Instances. Parent ref: group-v285292. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 625.466798] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9cdee956-d5ef-463b-92f5-4a4d4af0cd0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.476056] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Created folder: Instances in parent group-v285292. [ 625.476304] env[61006]: DEBUG oslo.service.loopingcall [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.476487] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 625.476688] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33db5409-6643-4291-97b4-145bb5370bc5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.492534] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 625.492534] env[61006]: value = "task-1336885" [ 625.492534] env[61006]: _type = "Task" [ 625.492534] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.500070] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336885, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.658554] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.948045] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.010765] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336885, 'name': CreateVM_Task, 'duration_secs': 0.27751} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.011210] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 626.013170] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.013170] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.013581] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 626.015923] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40ff7b74-d9c6-4905-82e3-9802c29c2931 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.023106] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 626.023106] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524ec9cb-b8a1-03d7-145f-514bc6b21e60" [ 626.023106] env[61006]: _type = "Task" [ 626.023106] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.029986] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524ec9cb-b8a1-03d7-145f-514bc6b21e60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.057327] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.379250] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d5484e-81b8-4f22-aa23-7e1735839461 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.388764] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5c703f-103c-4dab-a8cf-8e4ecd9f3571 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.421207] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87de8a7a-fabb-481d-b9a7-915498be47e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.428671] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9266839-5f68-4dd3-9ac7-9481d6191870 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.441756] env[61006]: DEBUG nova.compute.provider_tree [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.530435] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524ec9cb-b8a1-03d7-145f-514bc6b21e60, 'name': SearchDatastore_Task, 'duration_secs': 0.009777} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.530752] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.530984] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 626.531231] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.531373] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.531636] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 626.531865] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6defa28f-5f10-4cae-a0d9-f7ad7f82ad69 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.539178] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.539356] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 626.540076] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eed21fd-c372-4221-8862-e54ffae4f318 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.544781] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 626.544781] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f2facc-b25e-a099-6df2-359198e777fd" [ 626.544781] env[61006]: _type = "Task" [ 626.544781] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.553643] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f2facc-b25e-a099-6df2-359198e777fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.559348] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Releasing lock "refresh_cache-70e6f255-680f-4e92-a2fe-254127e70b77" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.559630] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.559874] env[61006]: DEBUG nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.560121] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.585754] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.945226] env[61006]: DEBUG nova.scheduler.client.report [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.055839] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f2facc-b25e-a099-6df2-359198e777fd, 'name': SearchDatastore_Task, 'duration_secs': 0.008538} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.056854] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-994816d1-27a0-4036-b18d-316ec7a1dcdb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.064262] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 627.064262] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267d336-c2ed-283a-69b2-729e6714dccc" [ 627.064262] env[61006]: _type = "Task" [ 627.064262] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.072650] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267d336-c2ed-283a-69b2-729e6714dccc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.092279] env[61006]: DEBUG nova.network.neutron [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.452120] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.452364] env[61006]: ERROR nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Traceback (most recent call last): [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.driver.spawn(context, instance, image_meta, [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] vm_ref = self.build_virtual_machine(instance, [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.452364] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] for vif in network_info: [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self._sync_wrapper(fn, *args, **kwargs) [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.wait() [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self[:] = self._gt.wait() [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self._exit_event.wait() [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] result = hub.switch() [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.452705] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return self.greenlet.switch() [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] result = function(*args, **kwargs) [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] return func(*args, **kwargs) [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise e [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] nwinfo = self.network_api.allocate_for_instance( [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] created_port_ids = self._update_ports_for_instance( [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] with excutils.save_and_reraise_exception(): [ 627.453373] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] self.force_reraise() [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise self.value [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] updated_port = self._update_port( [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] _ensure_no_port_binding_failure(port) [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] raise exception.PortBindingFailed(port_id=port['id']) [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] nova.exception.PortBindingFailed: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. [ 627.453728] env[61006]: ERROR nova.compute.manager [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] [ 627.454052] env[61006]: DEBUG nova.compute.utils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.455672] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.758s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.459987] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Build of instance f26bffa0-31b3-4a12-aa1a-3440032bd355 was re-scheduled: Binding failed for port 53c140f8-1546-438b-9f73-8d985e0a8065, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.459987] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.459987] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.459987] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.460304] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.575816] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267d336-c2ed-283a-69b2-729e6714dccc, 'name': SearchDatastore_Task, 'duration_secs': 0.009471} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.576227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.576269] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6642ffcb-cb01-4e38-a27c-bf4e4c938a17/6642ffcb-cb01-4e38-a27c-bf4e4c938a17.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 627.576517] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70e329b3-e834-46d1-85a2-10214aa5452c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.583081] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 627.583081] env[61006]: value = "task-1336886" [ 627.583081] env[61006]: _type = "Task" [ 627.583081] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.590785] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.593368] env[61006]: INFO nova.compute.manager [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] [instance: 70e6f255-680f-4e92-a2fe-254127e70b77] Took 1.03 seconds to deallocate network for instance. [ 627.991023] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.094240] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471828} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.094492] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6642ffcb-cb01-4e38-a27c-bf4e4c938a17/6642ffcb-cb01-4e38-a27c-bf4e4c938a17.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 628.094733] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 628.094989] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5ba4c44-de16-42b6-8059-39c399108d03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.106693] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 628.106693] env[61006]: value = "task-1336887" [ 628.106693] env[61006]: _type = "Task" [ 628.106693] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.114886] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336887, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.120401] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.406081] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3045897-024b-437a-899e-99515b871bf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.415067] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b28775-19a9-40ab-bb81-645bfb2963da {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.443997] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd606bd6-d946-4ffb-a6be-bd8361d03cb3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.455583] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec5d1b7-df74-4315-9c92-f801aaf7bab7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.467699] env[61006]: DEBUG nova.compute.provider_tree [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.616538] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336887, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067998} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.616833] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.618010] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdd0c7b-f2b2-4379-a2b0-718223785387 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.622504] env[61006]: INFO nova.scheduler.client.report [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Deleted allocations for instance 70e6f255-680f-4e92-a2fe-254127e70b77 [ 628.627873] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-f26bffa0-31b3-4a12-aa1a-3440032bd355" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.628091] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.628256] env[61006]: DEBUG nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.628417] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.647192] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 6642ffcb-cb01-4e38-a27c-bf4e4c938a17/6642ffcb-cb01-4e38-a27c-bf4e4c938a17.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.648109] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58f7e1c6-621a-4b95-97cf-d41e3d2b00f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.663971] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.670251] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 628.670251] env[61006]: value = "task-1336888" [ 628.670251] env[61006]: _type = "Task" [ 628.670251] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.679043] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336888, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.970686] env[61006]: DEBUG nova.scheduler.client.report [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.148907] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a281550b-2125-4ac9-bb7e-11530574a830 tempest-AttachInterfacesV270Test-947354748 tempest-AttachInterfacesV270Test-947354748-project-member] Lock "70e6f255-680f-4e92-a2fe-254127e70b77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.617s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.166770] env[61006]: DEBUG nova.network.neutron [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.182150] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336888, 'name': ReconfigVM_Task, 'duration_secs': 0.303636} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.182433] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 6642ffcb-cb01-4e38-a27c-bf4e4c938a17/6642ffcb-cb01-4e38-a27c-bf4e4c938a17.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.183361] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-343209cf-1510-41a0-827d-cd5b2363da17 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.189572] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 629.189572] env[61006]: value = "task-1336889" [ 629.189572] env[61006]: _type = "Task" [ 629.189572] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.198454] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336889, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.475477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.476147] env[61006]: ERROR nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Traceback (most recent call last): [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.driver.spawn(context, instance, image_meta, [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] vm_ref = self.build_virtual_machine(instance, [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.476147] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] for vif in network_info: [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return self._sync_wrapper(fn, *args, **kwargs) [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.wait() [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self[:] = self._gt.wait() [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return self._exit_event.wait() [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] current.throw(*self._exc) [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.476473] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] result = function(*args, **kwargs) [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] return func(*args, **kwargs) [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise e [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] nwinfo = self.network_api.allocate_for_instance( [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] created_port_ids = self._update_ports_for_instance( [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] with excutils.save_and_reraise_exception(): [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] self.force_reraise() [ 629.476883] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise self.value [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] updated_port = self._update_port( [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] _ensure_no_port_binding_failure(port) [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] raise exception.PortBindingFailed(port_id=port['id']) [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] nova.exception.PortBindingFailed: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. [ 629.477328] env[61006]: ERROR nova.compute.manager [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] [ 629.477328] env[61006]: DEBUG nova.compute.utils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.478302] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.268s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.479778] env[61006]: INFO nova.compute.claims [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.484540] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Build of instance e43ab4da-0b5d-4bde-a027-da603152bd72 was re-scheduled: Binding failed for port 1c0e2aed-aad5-49e2-8346-dbd2321c27f4, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.485019] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.485267] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.485412] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.485566] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.651729] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.669824] env[61006]: INFO nova.compute.manager [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: f26bffa0-31b3-4a12-aa1a-3440032bd355] Took 1.04 seconds to deallocate network for instance. [ 629.700786] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336889, 'name': Rename_Task, 'duration_secs': 0.172569} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.701685] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.701685] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fdc65ba1-aacc-45cc-9c08-770c66b608e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.708040] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 629.708040] env[61006]: value = "task-1336890" [ 629.708040] env[61006]: _type = "Task" [ 629.708040] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.717716] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336890, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.005769] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.109031] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.178559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.219500] env[61006]: DEBUG oslo_vmware.api [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336890, 'name': PowerOnVM_Task, 'duration_secs': 0.451546} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.219836] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.220140] env[61006]: INFO nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Took 4.85 seconds to spawn the instance on the hypervisor. [ 630.220359] env[61006]: DEBUG nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.221271] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8850318f-5fd6-4fc7-a087-c1240e1ff8f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.612180] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-e43ab4da-0b5d-4bde-a027-da603152bd72" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.612395] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 630.612572] env[61006]: DEBUG nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.612737] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.635084] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.702790] env[61006]: INFO nova.scheduler.client.report [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance f26bffa0-31b3-4a12-aa1a-3440032bd355 [ 630.738970] env[61006]: INFO nova.compute.manager [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Took 33.57 seconds to build instance. [ 630.951274] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48816c1-ee43-4521-9085-0ae18b9dd7eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.958219] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00d6178-baac-461f-8a30-98ecfba77e3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.990722] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76498b53-3a8b-4f1c-958f-381f5353b25f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.000211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1801e1a6-c15a-4c48-bc8d-c375d0650faa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.018660] env[61006]: DEBUG nova.compute.provider_tree [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.137344] env[61006]: DEBUG nova.network.neutron [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.212189] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f4ce4309-dcbf-4a59-bab6-b27cb59bb58a tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "f26bffa0-31b3-4a12-aa1a-3440032bd355" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.602s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.239321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d05a3f9a-b518-49bb-a456-1dc205b21572 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.692s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.525668] env[61006]: DEBUG nova.scheduler.client.report [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.648028] env[61006]: INFO nova.compute.manager [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: e43ab4da-0b5d-4bde-a027-da603152bd72] Took 1.03 seconds to deallocate network for instance. [ 631.721141] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.742250] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.031573] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.032192] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.035785] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.860s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.037187] env[61006]: INFO nova.compute.claims [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.243969] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.270697] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.542720] env[61006]: DEBUG nova.compute.utils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.547477] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.547477] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 632.607638] env[61006]: DEBUG nova.policy [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82aea0ad3a245af99868fc53a0b04f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd26132a67fc1412c938a36094a34a433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.687074] env[61006]: INFO nova.scheduler.client.report [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted allocations for instance e43ab4da-0b5d-4bde-a027-da603152bd72 [ 632.925037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquiring lock "dac63cd3-1788-4e45-b58a-9cd22038dd30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.925340] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "dac63cd3-1788-4e45-b58a-9cd22038dd30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.047201] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 633.193366] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Successfully created port: f7794882-a3ef-42b9-87ea-1ec4119afa11 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.198751] env[61006]: DEBUG oslo_concurrency.lockutils [None req-988e3096-2aeb-4e51-a7ed-43ce2a2d2363 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "e43ab4da-0b5d-4bde-a027-da603152bd72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.497s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.548252] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e087571-3184-4e2e-9ec5-f6276e1d697d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.559235] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c723de3-3cea-4795-91fb-c6e5fd1379c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.595949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fa347d-43d9-4227-a378-c612f08059fc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.603393] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8969fc4a-795a-4a43-ba69-2c56c42d6e96 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.619043] env[61006]: DEBUG nova.compute.provider_tree [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.706304] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.060473] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.097074] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.097991] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.097991] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.097991] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.098144] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.098295] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.098569] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.098657] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.098819] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.098975] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.099159] env[61006]: DEBUG nova.virt.hardware [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.100890] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "6ee89cb9-ef41-4c85-98d2-5b0190568efe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.100890] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "6ee89cb9-ef41-4c85-98d2-5b0190568efe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.100890] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af6bf99-239e-4197-bfac-21d2d1f2f916 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.109593] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c555779-4907-4ce2-9d7c-2fe0d7a8bd88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.123285] env[61006]: DEBUG nova.scheduler.client.report [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.229408] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.425353] env[61006]: DEBUG nova.compute.manager [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Received event network-changed-f7794882-a3ef-42b9-87ea-1ec4119afa11 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.425569] env[61006]: DEBUG nova.compute.manager [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Refreshing instance network info cache due to event network-changed-f7794882-a3ef-42b9-87ea-1ec4119afa11. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.425781] env[61006]: DEBUG oslo_concurrency.lockutils [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] Acquiring lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.425920] env[61006]: DEBUG oslo_concurrency.lockutils [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] Acquired lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.429287] env[61006]: DEBUG nova.network.neutron [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Refreshing network info cache for port f7794882-a3ef-42b9-87ea-1ec4119afa11 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.616434] env[61006]: ERROR nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 634.616434] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.616434] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.616434] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.616434] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.616434] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.616434] env[61006]: ERROR nova.compute.manager raise self.value [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.616434] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.616434] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.616434] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.617134] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.617134] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.617134] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 634.617134] env[61006]: ERROR nova.compute.manager [ 634.617134] env[61006]: Traceback (most recent call last): [ 634.617134] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.617134] env[61006]: listener.cb(fileno) [ 634.617134] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.617134] env[61006]: result = function(*args, **kwargs) [ 634.617134] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.617134] env[61006]: return func(*args, **kwargs) [ 634.617134] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.617134] env[61006]: raise e [ 634.617134] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.617134] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 634.617134] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.617134] env[61006]: created_port_ids = self._update_ports_for_instance( [ 634.617134] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.617134] env[61006]: with excutils.save_and_reraise_exception(): [ 634.617134] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.617134] env[61006]: self.force_reraise() [ 634.617134] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.617134] env[61006]: raise self.value [ 634.617134] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.617134] env[61006]: updated_port = self._update_port( [ 634.617134] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.617134] env[61006]: _ensure_no_port_binding_failure(port) [ 634.617134] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.617134] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.617821] env[61006]: nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 634.617821] env[61006]: Removing descriptor: 17 [ 634.617821] env[61006]: ERROR nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Traceback (most recent call last): [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] yield resources [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.driver.spawn(context, instance, image_meta, [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.617821] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] vm_ref = self.build_virtual_machine(instance, [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] for vif in network_info: [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self._sync_wrapper(fn, *args, **kwargs) [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.wait() [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self[:] = self._gt.wait() [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self._exit_event.wait() [ 634.618138] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] result = hub.switch() [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self.greenlet.switch() [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] result = function(*args, **kwargs) [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return func(*args, **kwargs) [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise e [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] nwinfo = self.network_api.allocate_for_instance( [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.618470] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] created_port_ids = self._update_ports_for_instance( [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] with excutils.save_and_reraise_exception(): [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.force_reraise() [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise self.value [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] updated_port = self._update_port( [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] _ensure_no_port_binding_failure(port) [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.618789] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise exception.PortBindingFailed(port_id=port['id']) [ 634.619092] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 634.619092] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] [ 634.619092] env[61006]: INFO nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Terminating instance [ 634.619916] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.628161] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.592s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.628630] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.632086] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.836s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.633711] env[61006]: INFO nova.compute.claims [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.961751] env[61006]: DEBUG nova.network.neutron [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.138192] env[61006]: DEBUG nova.compute.utils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 635.141691] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 635.141866] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.165063] env[61006]: DEBUG nova.network.neutron [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.252231] env[61006]: DEBUG nova.policy [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6c5fe37a73b40b48e7cbb4f4eaa3d74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '064b30c918ac499e9838488a08256d29', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.642848] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.667011] env[61006]: DEBUG oslo_concurrency.lockutils [req-f9569918-26c7-4dea-890c-ce79db2523ec req-5797142e-0b3e-473f-915f-bedbf4114768 service nova] Releasing lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.667429] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.667642] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.896559] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Successfully created port: bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.210933] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.242616] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942c6bae-7b41-467a-9b42-7f342a7e7734 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.250375] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c26d91a-1126-417c-a63c-173c25351e20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.284987] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afff264-9d5a-43d2-b4fd-bb005014c088 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.292932] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b7cb3b-5746-4330-82a0-10b306671f14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.306822] env[61006]: DEBUG nova.compute.provider_tree [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.598272] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.657038] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.694069] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.694069] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.694069] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.694272] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.694272] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.694272] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.694272] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.694376] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.694528] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.694691] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.694856] env[61006]: DEBUG nova.virt.hardware [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.695728] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6936dba7-11da-42f6-9f53-2a51d99b77c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.704091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b0dd28-265a-4d96-bea7-898f2089db20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.757766] env[61006]: DEBUG nova.compute.manager [req-ce3545c3-4ab0-43fc-b475-6b308d19bd8b req-2eafc5dd-bcdf-4f50-8ce4-0a1974351bec service nova] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Received event network-vif-deleted-f7794882-a3ef-42b9-87ea-1ec4119afa11 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.811768] env[61006]: DEBUG nova.scheduler.client.report [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.102403] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.102828] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.103023] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.103319] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a085238-ddd6-478e-a714-b16749662226 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.113073] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740e0a42-71e8-4fa6-b498-e3977c94f70c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.137163] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1fdae918-be3c-441d-8fee-ff5a2a32373c could not be found. [ 637.137258] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.137401] env[61006]: INFO nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 637.137654] env[61006]: DEBUG oslo.service.loopingcall [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.137905] env[61006]: DEBUG nova.compute.manager [-] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.138013] env[61006]: DEBUG nova.network.neutron [-] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.154757] env[61006]: DEBUG nova.network.neutron [-] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.317075] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.318107] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.321444] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.823s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.324210] env[61006]: INFO nova.compute.claims [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.450784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.451381] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.658225] env[61006]: DEBUG nova.network.neutron [-] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.680727] env[61006]: ERROR nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 637.680727] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.680727] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.680727] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.680727] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.680727] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.680727] env[61006]: ERROR nova.compute.manager raise self.value [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.680727] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.680727] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.680727] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.681163] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.681163] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.681163] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 637.681163] env[61006]: ERROR nova.compute.manager [ 637.681163] env[61006]: Traceback (most recent call last): [ 637.681163] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.681163] env[61006]: listener.cb(fileno) [ 637.681163] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.681163] env[61006]: result = function(*args, **kwargs) [ 637.681163] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.681163] env[61006]: return func(*args, **kwargs) [ 637.681163] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.681163] env[61006]: raise e [ 637.681163] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.681163] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 637.681163] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.681163] env[61006]: created_port_ids = self._update_ports_for_instance( [ 637.681163] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.681163] env[61006]: with excutils.save_and_reraise_exception(): [ 637.681163] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.681163] env[61006]: self.force_reraise() [ 637.681163] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.681163] env[61006]: raise self.value [ 637.681163] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.681163] env[61006]: updated_port = self._update_port( [ 637.681163] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.681163] env[61006]: _ensure_no_port_binding_failure(port) [ 637.681163] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.681163] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.681823] env[61006]: nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 637.681823] env[61006]: Removing descriptor: 20 [ 637.681823] env[61006]: ERROR nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Traceback (most recent call last): [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] yield resources [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.driver.spawn(context, instance, image_meta, [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.681823] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] vm_ref = self.build_virtual_machine(instance, [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] for vif in network_info: [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self._sync_wrapper(fn, *args, **kwargs) [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.wait() [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self[:] = self._gt.wait() [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self._exit_event.wait() [ 637.682142] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] result = hub.switch() [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self.greenlet.switch() [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] result = function(*args, **kwargs) [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return func(*args, **kwargs) [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise e [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] nwinfo = self.network_api.allocate_for_instance( [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.682488] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] created_port_ids = self._update_ports_for_instance( [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] with excutils.save_and_reraise_exception(): [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.force_reraise() [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise self.value [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] updated_port = self._update_port( [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] _ensure_no_port_binding_failure(port) [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.682833] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise exception.PortBindingFailed(port_id=port['id']) [ 637.683122] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 637.683122] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] [ 637.683122] env[61006]: INFO nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Terminating instance [ 637.684714] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.684882] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.685053] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.829595] env[61006]: DEBUG nova.compute.utils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.832806] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 638.161044] env[61006]: INFO nova.compute.manager [-] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Took 1.02 seconds to deallocate network for instance. [ 638.165160] env[61006]: DEBUG nova.compute.claims [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.165160] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.203417] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.323562] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.336023] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.744698] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e017d6-a02b-48fc-9ab0-ad3043cb11ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.750356] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef97de37-8240-4520-b8e0-f35ef49ab827 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.783190] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ae0d9b-c6cf-43f7-b753-e5f5ff787628 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.785226] env[61006]: DEBUG nova.compute.manager [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Received event network-changed-bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.785412] env[61006]: DEBUG nova.compute.manager [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Refreshing instance network info cache due to event network-changed-bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 638.785598] env[61006]: DEBUG oslo_concurrency.lockutils [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] Acquiring lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.791246] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8a3508-fe86-4111-acb2-e01af14950ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.804588] env[61006]: DEBUG nova.compute.provider_tree [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.826645] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.826997] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.827200] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.827496] env[61006]: DEBUG oslo_concurrency.lockutils [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] Acquired lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.827701] env[61006]: DEBUG nova.network.neutron [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Refreshing network info cache for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.828610] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c091bea2-8df2-4f88-a716-6871c187d305 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.837773] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce58a1c7-1d63-4a2c-846d-36f2315b567d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.861674] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e2ef686c-b517-4121-88f5-01cad6b960b9 could not be found. [ 638.861932] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.862129] env[61006]: INFO nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.862362] env[61006]: DEBUG oslo.service.loopingcall [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.862576] env[61006]: DEBUG nova.compute.manager [-] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.862671] env[61006]: DEBUG nova.network.neutron [-] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.878279] env[61006]: DEBUG nova.network.neutron [-] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.149155] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.149376] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.307996] env[61006]: DEBUG nova.scheduler.client.report [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.351527] env[61006]: DEBUG nova.network.neutron [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.354100] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.380770] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.380770] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.380927] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.381076] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.381190] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.381298] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.381508] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.381678] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.381950] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.382011] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.382187] env[61006]: DEBUG nova.virt.hardware [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.382606] env[61006]: DEBUG nova.network.neutron [-] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.384616] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130ba92b-0b00-44d1-9185-c4de7ef2b53b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.396227] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dc95d5-963e-4f88-842f-3b9c4ad9bb1e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.412610] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.418747] env[61006]: DEBUG oslo.service.loopingcall [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.421106] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.421341] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7170ee41-5b9f-4461-b34f-1de06199c656 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.440161] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.440161] env[61006]: value = "task-1336891" [ 639.440161] env[61006]: _type = "Task" [ 639.440161] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.447928] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336891, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.468295] env[61006]: DEBUG nova.network.neutron [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.815128] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.815128] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.816570] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.578s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.816738] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.816887] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 639.817192] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.220s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.820803] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5aa4ae-006c-452d-b361-347a4223542f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.829201] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2451ebc7-5d13-45c2-ab31-e5d7f4863f14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.842673] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249ff99e-9f11-4f94-97e7-4139c3cffa8c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.849785] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d06a776-5d3f-46d3-8afd-1b7618ec23bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.879988] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181449MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 639.880162] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.889326] env[61006]: INFO nova.compute.manager [-] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Took 1.03 seconds to deallocate network for instance. [ 639.891496] env[61006]: DEBUG nova.compute.claims [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.891662] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.950066] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336891, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.970852] env[61006]: DEBUG oslo_concurrency.lockutils [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] Releasing lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.971229] env[61006]: DEBUG nova.compute.manager [req-32eb40a8-f65d-4385-bfc3-cba8e4c829b6 req-bb8d333f-507c-414f-9e03-167db4a85214 service nova] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Received event network-vif-deleted-bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.321492] env[61006]: DEBUG nova.compute.utils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.323010] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.323236] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.376107] env[61006]: DEBUG nova.policy [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c86635fcebb342c68bbcc54a4122df23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3462ebe2b974e84ab820c1ea81dc945', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.451376] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336891, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.711127] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Successfully created port: d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.796718] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06c7f8a-2fca-4e48-aa3b-292e363299ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.904091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38674455-f333-45c1-b849-206afabdc257 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.904091] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.904091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc50d4f-5906-42d7-8569-950e69f14f7c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.904091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e382aa-9652-4980-9204-44ca1e5ad6b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.912879] env[61006]: DEBUG nova.compute.provider_tree [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.952306] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336891, 'name': CreateVM_Task, 'duration_secs': 1.251148} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.952499] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 640.952950] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.953166] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.953488] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 640.953750] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de3f539f-3076-4f3c-9dd9-2bffaf5fe962 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.959624] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 640.959624] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058b6-2540-7b07-2747-136b1e5f41bb" [ 640.959624] env[61006]: _type = "Task" [ 640.959624] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.970981] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058b6-2540-7b07-2747-136b1e5f41bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.416515] env[61006]: DEBUG nova.scheduler.client.report [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.471888] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058b6-2540-7b07-2747-136b1e5f41bb, 'name': SearchDatastore_Task, 'duration_secs': 0.02217} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.473042] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.473042] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.473042] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.473042] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.473042] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.473352] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ac0ab38-f881-4c42-a83e-19a7b00b7caa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.481985] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.482103] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 641.482737] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c74c4b5f-7dad-4888-a75d-842c47658902 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.488082] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 641.488082] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f1387-7fd1-5025-eadb-6308daad4cca" [ 641.488082] env[61006]: _type = "Task" [ 641.488082] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.495989] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f1387-7fd1-5025-eadb-6308daad4cca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.700670] env[61006]: DEBUG nova.compute.manager [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Received event network-changed-d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.700900] env[61006]: DEBUG nova.compute.manager [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Refreshing instance network info cache due to event network-changed-d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.701092] env[61006]: DEBUG oslo_concurrency.lockutils [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] Acquiring lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.701207] env[61006]: DEBUG oslo_concurrency.lockutils [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] Acquired lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.701365] env[61006]: DEBUG nova.network.neutron [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Refreshing network info cache for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.878298] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.906265] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.906571] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.906725] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.906941] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.907181] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.907345] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.907707] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.907976] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.908365] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.908590] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.908883] env[61006]: DEBUG nova.virt.hardware [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.910456] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295ceb9a-ad15-461f-bc05-2a4af2801be9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.916157] env[61006]: ERROR nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 641.916157] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.916157] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.916157] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.916157] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.916157] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.916157] env[61006]: ERROR nova.compute.manager raise self.value [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.916157] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.916157] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.916157] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.916635] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.916635] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.916635] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 641.916635] env[61006]: ERROR nova.compute.manager [ 641.916635] env[61006]: Traceback (most recent call last): [ 641.916635] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.916635] env[61006]: listener.cb(fileno) [ 641.916635] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.916635] env[61006]: result = function(*args, **kwargs) [ 641.916635] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.916635] env[61006]: return func(*args, **kwargs) [ 641.916635] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.916635] env[61006]: raise e [ 641.916635] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.916635] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 641.916635] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.916635] env[61006]: created_port_ids = self._update_ports_for_instance( [ 641.916635] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.916635] env[61006]: with excutils.save_and_reraise_exception(): [ 641.916635] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.916635] env[61006]: self.force_reraise() [ 641.916635] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.916635] env[61006]: raise self.value [ 641.916635] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.916635] env[61006]: updated_port = self._update_port( [ 641.916635] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.916635] env[61006]: _ensure_no_port_binding_failure(port) [ 641.916635] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.916635] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.917559] env[61006]: nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 641.917559] env[61006]: Removing descriptor: 20 [ 641.923017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.106s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.923946] env[61006]: ERROR nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Traceback (most recent call last): [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.driver.spawn(context, instance, image_meta, [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] vm_ref = self.build_virtual_machine(instance, [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.923946] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] for vif in network_info: [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self._sync_wrapper(fn, *args, **kwargs) [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.wait() [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self[:] = self._gt.wait() [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self._exit_event.wait() [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] result = hub.switch() [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.924593] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return self.greenlet.switch() [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] result = function(*args, **kwargs) [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] return func(*args, **kwargs) [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise e [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] nwinfo = self.network_api.allocate_for_instance( [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] created_port_ids = self._update_ports_for_instance( [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] with excutils.save_and_reraise_exception(): [ 641.924927] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] self.force_reraise() [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise self.value [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] updated_port = self._update_port( [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] _ensure_no_port_binding_failure(port) [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] raise exception.PortBindingFailed(port_id=port['id']) [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] nova.exception.PortBindingFailed: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. [ 641.925296] env[61006]: ERROR nova.compute.manager [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] [ 641.925624] env[61006]: DEBUG nova.compute.utils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.926743] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.270s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.928994] env[61006]: INFO nova.compute.claims [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.934472] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad121ca-6988-47ed-b5f9-b35643d165b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.942215] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Build of instance de11f24e-7aab-4a68-a099-c5a8ca32259b was re-scheduled: Binding failed for port bebca07e-1a36-4b2c-ad30-ba4a3d623b58, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.943371] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.943965] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquiring lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.944384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Acquired lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.944648] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.957269] env[61006]: ERROR nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] Traceback (most recent call last): [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] yield resources [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.driver.spawn(context, instance, image_meta, [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] vm_ref = self.build_virtual_machine(instance, [ 641.957269] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] for vif in network_info: [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return self._sync_wrapper(fn, *args, **kwargs) [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.wait() [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self[:] = self._gt.wait() [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return self._exit_event.wait() [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.957626] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] current.throw(*self._exc) [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] result = function(*args, **kwargs) [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return func(*args, **kwargs) [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise e [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] nwinfo = self.network_api.allocate_for_instance( [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] created_port_ids = self._update_ports_for_instance( [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] with excutils.save_and_reraise_exception(): [ 641.958076] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.force_reraise() [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise self.value [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] updated_port = self._update_port( [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] _ensure_no_port_binding_failure(port) [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise exception.PortBindingFailed(port_id=port['id']) [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 641.958406] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] [ 641.958406] env[61006]: INFO nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Terminating instance [ 641.959929] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquiring lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.997873] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f1387-7fd1-5025-eadb-6308daad4cca, 'name': SearchDatastore_Task, 'duration_secs': 0.012616} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.998665] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6408737-82db-4e68-85f5-181ef7bd7d14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.003843] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 642.003843] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2e163-3dde-8d8c-0963-fade0a5f55ab" [ 642.003843] env[61006]: _type = "Task" [ 642.003843] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.011760] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2e163-3dde-8d8c-0963-fade0a5f55ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.226290] env[61006]: DEBUG nova.network.neutron [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.316255] env[61006]: DEBUG nova.network.neutron [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.472030] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.513765] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2e163-3dde-8d8c-0963-fade0a5f55ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008708} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.514052] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.514319] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 642.514701] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bac9521b-d456-4339-a792-41eb68991a8a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.522185] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 642.522185] env[61006]: value = "task-1336892" [ 642.522185] env[61006]: _type = "Task" [ 642.522185] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.530164] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336892, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.548255] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.819716] env[61006]: DEBUG oslo_concurrency.lockutils [req-bb7976e8-743b-46da-a4cb-3d1887309624 req-6cadb38d-f097-4df3-a540-061867980e2e service nova] Releasing lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.820222] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquired lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.820418] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.031394] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336892, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468547} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.031668] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 643.032059] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 643.032243] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82f509c9-b465-4068-9bf5-7aa3bd87e4db {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.040024] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 643.040024] env[61006]: value = "task-1336893" [ 643.040024] env[61006]: _type = "Task" [ 643.040024] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.046293] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.051079] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Releasing lock "refresh_cache-de11f24e-7aab-4a68-a099-c5a8ca32259b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.051350] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.051519] env[61006]: DEBUG nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.051708] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.075530] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.347371] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.424042] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.436568] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e6ac50-c839-4c27-809f-cf8bf6bf8986 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.444836] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebb5cf3-3ac9-4fdb-88e5-a23d5cab5100 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.475666] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7610d9ba-5a57-4a9e-a2f9-d5be9608b026 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.483378] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983aa1a8-8a20-48db-8ac9-0d85ea2f0ccf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.497304] env[61006]: DEBUG nova.compute.provider_tree [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.547676] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066976} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.547942] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.548772] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64cded5-e2bd-40bb-8455-612f3ce04015 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.568168] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.568479] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de7cec4c-f499-4f1b-ad83-400798bd426a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.583489] env[61006]: DEBUG nova.network.neutron [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.590617] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 643.590617] env[61006]: value = "task-1336894" [ 643.590617] env[61006]: _type = "Task" [ 643.590617] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.599767] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.724091] env[61006]: DEBUG nova.compute.manager [req-e204580d-f465-40c6-a98b-5725f6321668 req-77e60812-3414-4dfa-8777-131e4ad83521 service nova] [instance: aca01468-634e-4013-8a96-c7139f007c60] Received event network-vif-deleted-d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.927077] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Releasing lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.927529] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.927747] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.928062] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f53ebad5-5dc8-46fc-b10e-1c0365601030 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.938897] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9075c877-eaba-4698-8639-190dfffde3cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.961208] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aca01468-634e-4013-8a96-c7139f007c60 could not be found. [ 643.961498] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.961722] env[61006]: INFO nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.962035] env[61006]: DEBUG oslo.service.loopingcall [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.962312] env[61006]: DEBUG nova.compute.manager [-] [instance: aca01468-634e-4013-8a96-c7139f007c60] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.962414] env[61006]: DEBUG nova.network.neutron [-] [instance: aca01468-634e-4013-8a96-c7139f007c60] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.977681] env[61006]: DEBUG nova.network.neutron [-] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.000516] env[61006]: DEBUG nova.scheduler.client.report [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.087030] env[61006]: INFO nova.compute.manager [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] [instance: de11f24e-7aab-4a68-a099-c5a8ca32259b] Took 1.03 seconds to deallocate network for instance. [ 644.100504] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336894, 'name': ReconfigVM_Task, 'duration_secs': 0.266233} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.101110] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 644.101939] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca083660-438e-4e97-b746-083ffcebc5be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.107937] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 644.107937] env[61006]: value = "task-1336895" [ 644.107937] env[61006]: _type = "Task" [ 644.107937] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.115508] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336895, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.479915] env[61006]: DEBUG nova.network.neutron [-] [instance: aca01468-634e-4013-8a96-c7139f007c60] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.505453] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.505953] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.508396] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.330s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.509783] env[61006]: INFO nova.compute.claims [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.618687] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336895, 'name': Rename_Task, 'duration_secs': 0.130524} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.619178] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.619293] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef3708b8-8611-4a16-8026-d3f8272bd3ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.626178] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 644.626178] env[61006]: value = "task-1336896" [ 644.626178] env[61006]: _type = "Task" [ 644.626178] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.634494] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.982590] env[61006]: INFO nova.compute.manager [-] [instance: aca01468-634e-4013-8a96-c7139f007c60] Took 1.02 seconds to deallocate network for instance. [ 644.985519] env[61006]: DEBUG nova.compute.claims [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.985703] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.014457] env[61006]: DEBUG nova.compute.utils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.017657] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.017850] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.056650] env[61006]: DEBUG nova.policy [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '292c8fd5d18e4779a07afa3ced39d80b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4e5ebdeaa574dd981e388a7f643a189', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.117342] env[61006]: INFO nova.scheduler.client.report [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Deleted allocations for instance de11f24e-7aab-4a68-a099-c5a8ca32259b [ 645.136822] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336896, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.374122] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Successfully created port: a122f146-2428-4908-938f-268937fa692f {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.523901] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.625801] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d011b1a6-4470-4114-a4c5-7499fe0156df tempest-DeleteServersAdminTestJSON-1643472066 tempest-DeleteServersAdminTestJSON-1643472066-project-member] Lock "de11f24e-7aab-4a68-a099-c5a8ca32259b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.702s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.636510] env[61006]: DEBUG oslo_vmware.api [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336896, 'name': PowerOnVM_Task, 'duration_secs': 0.786031} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.639049] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.639406] env[61006]: INFO nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Took 6.28 seconds to spawn the instance on the hypervisor. [ 645.639603] env[61006]: DEBUG nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 645.640663] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b208c29-88a3-41ee-aa3a-3e7e8c186fd3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.981921] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d207aa4-887d-4777-8dfd-fb2c65be216e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.992017] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b9f719-d32b-4919-8d27-9bc76bc233f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.023552] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d26db59-02f2-4915-8ed6-5ae0f48d8619 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.038615] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8193ff-3189-4de0-a496-a55f67dc7e95 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.055015] env[61006]: DEBUG nova.compute.provider_tree [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.132903] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.168139] env[61006]: INFO nova.compute.manager [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Took 29.40 seconds to build instance. [ 646.367536] env[61006]: DEBUG nova.compute.manager [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Received event network-changed-a122f146-2428-4908-938f-268937fa692f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.367758] env[61006]: DEBUG nova.compute.manager [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Refreshing instance network info cache due to event network-changed-a122f146-2428-4908-938f-268937fa692f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.368149] env[61006]: DEBUG oslo_concurrency.lockutils [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] Acquiring lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.368274] env[61006]: DEBUG oslo_concurrency.lockutils [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] Acquired lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.368467] env[61006]: DEBUG nova.network.neutron [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Refreshing network info cache for port a122f146-2428-4908-938f-268937fa692f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.537014] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.549517] env[61006]: ERROR nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 646.549517] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.549517] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.549517] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.549517] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.549517] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.549517] env[61006]: ERROR nova.compute.manager raise self.value [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.549517] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.549517] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.549517] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.550012] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.550012] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.550012] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 646.550012] env[61006]: ERROR nova.compute.manager [ 646.550012] env[61006]: Traceback (most recent call last): [ 646.550012] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.550012] env[61006]: listener.cb(fileno) [ 646.550012] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.550012] env[61006]: result = function(*args, **kwargs) [ 646.550012] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.550012] env[61006]: return func(*args, **kwargs) [ 646.550012] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.550012] env[61006]: raise e [ 646.550012] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.550012] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 646.550012] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.550012] env[61006]: created_port_ids = self._update_ports_for_instance( [ 646.550012] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.550012] env[61006]: with excutils.save_and_reraise_exception(): [ 646.550012] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.550012] env[61006]: self.force_reraise() [ 646.550012] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.550012] env[61006]: raise self.value [ 646.550012] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.550012] env[61006]: updated_port = self._update_port( [ 646.550012] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.550012] env[61006]: _ensure_no_port_binding_failure(port) [ 646.550012] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.550012] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.550675] env[61006]: nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 646.550675] env[61006]: Removing descriptor: 20 [ 646.559300] env[61006]: DEBUG nova.scheduler.client.report [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.574502] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.574741] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.574895] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.575084] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.575232] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.575365] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.575565] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.575722] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.575880] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.576049] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.576243] env[61006]: DEBUG nova.virt.hardware [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.577126] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ec75ac-3837-4708-ac47-03ec29c5ef9c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.586815] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9039382c-0a72-400d-a579-923f66d76982 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.601435] env[61006]: ERROR nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Traceback (most recent call last): [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] yield resources [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.driver.spawn(context, instance, image_meta, [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] vm_ref = self.build_virtual_machine(instance, [ 646.601435] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] for vif in network_info: [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return self._sync_wrapper(fn, *args, **kwargs) [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.wait() [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self[:] = self._gt.wait() [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return self._exit_event.wait() [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.601801] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] current.throw(*self._exc) [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] result = function(*args, **kwargs) [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return func(*args, **kwargs) [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise e [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] nwinfo = self.network_api.allocate_for_instance( [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] created_port_ids = self._update_ports_for_instance( [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] with excutils.save_and_reraise_exception(): [ 646.602164] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.force_reraise() [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise self.value [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] updated_port = self._update_port( [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] _ensure_no_port_binding_failure(port) [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise exception.PortBindingFailed(port_id=port['id']) [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 646.602518] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] [ 646.602518] env[61006]: INFO nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Terminating instance [ 646.603811] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquiring lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.658538] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.669207] env[61006]: DEBUG oslo_concurrency.lockutils [None req-02f02620-7362-43fa-ae41-181695b8b015 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.782s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.887259] env[61006]: DEBUG nova.network.neutron [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.971027] env[61006]: DEBUG nova.network.neutron [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.071023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.071023] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.072983] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.829s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.074693] env[61006]: INFO nova.compute.claims [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.173247] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.192919] env[61006]: INFO nova.compute.manager [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Rebuilding instance [ 647.243247] env[61006]: DEBUG nova.compute.manager [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 647.246718] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01e4fad-4e31-48be-92e6-9e55b80aeeed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.473896] env[61006]: DEBUG oslo_concurrency.lockutils [req-54d9742e-b054-4414-98a3-9386125c81c8 req-69dad523-ec5b-41ad-833c-fa8196e7f25e service nova] Releasing lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.474350] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquired lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.474779] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.577289] env[61006]: DEBUG nova.compute.utils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.577289] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 647.577289] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.665957] env[61006]: DEBUG nova.policy [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab0648a37b8a4aaba937edced69c20f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86d059b3ce274cb7a16c9e4bb4f843c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.692894] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.756431] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 647.756763] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e06ec4e-0a40-4e70-91a2-7fcd35daef4f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.764808] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 647.764808] env[61006]: value = "task-1336897" [ 647.764808] env[61006]: _type = "Task" [ 647.764808] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.775136] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.002886] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.040945] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Successfully created port: 72a7c63c-3585-4b4a-bf51-f6ebb89a232f {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.080242] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.155630] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.274606] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336897, 'name': PowerOffVM_Task, 'duration_secs': 0.211292} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.277217] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 648.277487] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.279022] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06ad0f2-3e58-43fa-959f-3e449b251254 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.285752] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 648.285994] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0126d860-b48a-46da-b945-dec38c6af2b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.321009] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 648.321448] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 648.321764] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleting the datastore file [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 648.322316] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b2299a1-299a-4e57-8086-c1d3999d9d28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.332905] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 648.332905] env[61006]: value = "task-1336899" [ 648.332905] env[61006]: _type = "Task" [ 648.332905] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.344333] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.486865] env[61006]: DEBUG nova.compute.manager [req-02ca11f8-ff73-4343-ac08-48b32fafd2f7 req-5f495cb8-a96d-4f31-aab5-cacfe2d4de9d service nova] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Received event network-vif-deleted-a122f146-2428-4908-938f-268937fa692f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.572373] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c319f65d-c61a-4136-b2ce-1db715639d21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.579700] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8ccb73-9f8c-4537-9989-6315a987a432 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.615173] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62de9c51-e398-4d0e-9134-dbb9b6782d4d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.623837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421ecb82-dd46-47bd-9264-3029b20de89d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.640047] env[61006]: DEBUG nova.compute.provider_tree [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.660084] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Releasing lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.660569] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.660689] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.661205] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-084d2d33-346b-4cbc-a832-eec17524e687 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.669576] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7143ec-a2cd-489f-bb4f-e6f963d70fc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.692369] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e42de415-1d9d-4931-a5b7-97112af42fd1 could not be found. [ 648.692583] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.692761] env[61006]: INFO nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 648.692995] env[61006]: DEBUG oslo.service.loopingcall [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.693255] env[61006]: DEBUG nova.compute.manager [-] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.693348] env[61006]: DEBUG nova.network.neutron [-] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.714669] env[61006]: DEBUG nova.network.neutron [-] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.841903] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111489} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.842322] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 648.842522] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 648.842698] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.116045] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.141554] env[61006]: ERROR nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 649.141554] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.141554] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.141554] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.141554] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.141554] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.141554] env[61006]: ERROR nova.compute.manager raise self.value [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.141554] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.141554] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.141554] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.142089] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.142089] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.142089] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 649.142089] env[61006]: ERROR nova.compute.manager [ 649.142089] env[61006]: Traceback (most recent call last): [ 649.142089] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.142089] env[61006]: listener.cb(fileno) [ 649.142089] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.142089] env[61006]: result = function(*args, **kwargs) [ 649.142089] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.142089] env[61006]: return func(*args, **kwargs) [ 649.142089] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.142089] env[61006]: raise e [ 649.142089] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.142089] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 649.142089] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.142089] env[61006]: created_port_ids = self._update_ports_for_instance( [ 649.142089] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.142089] env[61006]: with excutils.save_and_reraise_exception(): [ 649.142089] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.142089] env[61006]: self.force_reraise() [ 649.142089] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.142089] env[61006]: raise self.value [ 649.142089] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.142089] env[61006]: updated_port = self._update_port( [ 649.142089] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.142089] env[61006]: _ensure_no_port_binding_failure(port) [ 649.142089] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.142089] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.142973] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 649.142973] env[61006]: Removing descriptor: 20 [ 649.146419] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:09:18Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='802705af-e612-4a78-b3dc-b4aa0bae569b',id=34,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-442361725',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.146419] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.146419] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.146640] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.146640] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.146987] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.147368] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.147670] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.147965] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.148290] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.148607] env[61006]: DEBUG nova.virt.hardware [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.149708] env[61006]: DEBUG nova.scheduler.client.report [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.154239] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da8606f-9fff-4ddb-8dcf-c657dd85cfa8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.163476] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76154b4-e923-4d9f-a827-434981010987 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.178596] env[61006]: ERROR nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Traceback (most recent call last): [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] yield resources [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.driver.spawn(context, instance, image_meta, [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] vm_ref = self.build_virtual_machine(instance, [ 649.178596] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] for vif in network_info: [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return self._sync_wrapper(fn, *args, **kwargs) [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.wait() [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self[:] = self._gt.wait() [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return self._exit_event.wait() [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 649.178909] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] current.throw(*self._exc) [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] result = function(*args, **kwargs) [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return func(*args, **kwargs) [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise e [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] nwinfo = self.network_api.allocate_for_instance( [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] created_port_ids = self._update_ports_for_instance( [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] with excutils.save_and_reraise_exception(): [ 649.179346] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.force_reraise() [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise self.value [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] updated_port = self._update_port( [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] _ensure_no_port_binding_failure(port) [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise exception.PortBindingFailed(port_id=port['id']) [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 649.179671] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] [ 649.179671] env[61006]: INFO nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Terminating instance [ 649.181034] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.181034] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquired lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.181935] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.217932] env[61006]: DEBUG nova.network.neutron [-] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.659477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.660066] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.663198] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.392s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.664606] env[61006]: INFO nova.compute.claims [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.701168] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.720257] env[61006]: INFO nova.compute.manager [-] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Took 1.03 seconds to deallocate network for instance. [ 649.724627] env[61006]: DEBUG nova.compute.claims [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.724755] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.772599] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.880776] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.881041] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.881210] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.881388] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.881529] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.881668] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.881867] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.882035] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.882201] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.882359] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.882525] env[61006]: DEBUG nova.virt.hardware [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.883404] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857fd743-c0bf-4549-9336-1b4e77217371 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.893278] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5ac198-c135-425c-8e6c-47e6e5765b15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.906851] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 649.912535] env[61006]: DEBUG oslo.service.loopingcall [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.913156] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 649.913407] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ad416e4-4091-4f51-8271-aff6a8d4e2dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.929927] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 649.929927] env[61006]: value = "task-1336900" [ 649.929927] env[61006]: _type = "Task" [ 649.929927] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.937872] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336900, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.170680] env[61006]: DEBUG nova.compute.utils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.173675] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.173783] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.229656] env[61006]: DEBUG nova.policy [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdcaec59095a47e4acf3b832d7edf67f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b58e44ef36ba47f8ae64444ed6d0d964', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 650.274665] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Releasing lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.275110] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 650.275311] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.275612] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af075eb3-214f-4213-a4c5-de7fbf7569ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.284888] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0acc3d7-bcbd-4aa2-938b-ffd2943f7bdd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.308473] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cf4e241c-a4e3-438c-8059-c7f7cb41cf45 could not be found. [ 650.308611] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.309028] env[61006]: INFO nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Took 0.03 seconds to destroy the instance on the hypervisor. [ 650.309028] env[61006]: DEBUG oslo.service.loopingcall [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.309246] env[61006]: DEBUG nova.compute.manager [-] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 650.309335] env[61006]: DEBUG nova.network.neutron [-] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 650.337826] env[61006]: DEBUG nova.network.neutron [-] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.446613] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336900, 'name': CreateVM_Task, 'duration_secs': 0.253101} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.447395] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 650.450979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.450979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.450979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 650.450979] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddf6f10b-a6cf-4860-97ea-b3e442fdd898 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.455245] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 650.455245] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528c8a54-9f74-ff44-ecba-c5b7af2f39f9" [ 650.455245] env[61006]: _type = "Task" [ 650.455245] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.464079] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528c8a54-9f74-ff44-ecba-c5b7af2f39f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.519935] env[61006]: DEBUG nova.compute.manager [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Received event network-changed-72a7c63c-3585-4b4a-bf51-f6ebb89a232f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.520122] env[61006]: DEBUG nova.compute.manager [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Refreshing instance network info cache due to event network-changed-72a7c63c-3585-4b4a-bf51-f6ebb89a232f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.520322] env[61006]: DEBUG oslo_concurrency.lockutils [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] Acquiring lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.520454] env[61006]: DEBUG oslo_concurrency.lockutils [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] Acquired lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.520609] env[61006]: DEBUG nova.network.neutron [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Refreshing network info cache for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.637343] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Successfully created port: 60aca473-1610-4bac-a857-d394c905c592 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.674788] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.841956] env[61006]: DEBUG nova.network.neutron [-] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.969563] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528c8a54-9f74-ff44-ecba-c5b7af2f39f9, 'name': SearchDatastore_Task, 'duration_secs': 0.011366} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.969994] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.970477] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 650.970771] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.970971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.971250] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 650.973837] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e356d224-d1e0-4b39-b331-4f373d4066ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.982228] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 650.983136] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 650.983136] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55ba9de1-bd1f-4bec-9d3f-4b9684c13733 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.990122] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 650.990122] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5221b01e-75c7-58cb-060b-cbd0bde11b5e" [ 650.990122] env[61006]: _type = "Task" [ 650.990122] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.999105] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5221b01e-75c7-58cb-060b-cbd0bde11b5e, 'name': SearchDatastore_Task, 'duration_secs': 0.007635} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.999936] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd6be780-5e01-472c-a980-2a743c575e69 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.007159] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 651.007159] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c0aa-a141-06ae-8d32-0f41c5ed3d7e" [ 651.007159] env[61006]: _type = "Task" [ 651.007159] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.015461] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c0aa-a141-06ae-8d32-0f41c5ed3d7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.041266] env[61006]: DEBUG nova.network.neutron [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.126023] env[61006]: DEBUG nova.network.neutron [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.149424] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0f81a8-aa9e-476f-b683-73c711780837 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.158468] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d082510-4a45-48c3-ac40-af410ebe30a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.193284] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c916e13d-5856-4b48-ac03-4dbbd85fc1dc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.200891] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9ba732-80ed-4353-931e-a183f71d82ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.215827] env[61006]: DEBUG nova.compute.provider_tree [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.346132] env[61006]: INFO nova.compute.manager [-] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Took 1.04 seconds to deallocate network for instance. [ 651.348065] env[61006]: DEBUG nova.compute.claims [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 651.348224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.517563] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c0aa-a141-06ae-8d32-0f41c5ed3d7e, 'name': SearchDatastore_Task, 'duration_secs': 0.007604} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.521414] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.521696] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 651.521959] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f90a8e91-d2b6-4553-8015-3acdba1f453d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.529564] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 651.529564] env[61006]: value = "task-1336901" [ 651.529564] env[61006]: _type = "Task" [ 651.529564] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.537301] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.628596] env[61006]: DEBUG oslo_concurrency.lockutils [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] Releasing lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.628883] env[61006]: DEBUG nova.compute.manager [req-192ceb59-d054-4171-a8af-cecdd0d4dbc9 req-95fe497e-4ca9-422f-a29e-dbb963678828 service nova] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Received event network-vif-deleted-72a7c63c-3585-4b4a-bf51-f6ebb89a232f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.695557] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.723600] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.723600] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.723600] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.723812] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.723812] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.723904] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.724089] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.724219] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.724382] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.724538] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.724703] env[61006]: DEBUG nova.virt.hardware [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.725595] env[61006]: DEBUG nova.scheduler.client.report [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.729606] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917a1084-0915-45cb-9113-717540325702 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.733159] env[61006]: ERROR nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 651.733159] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.733159] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.733159] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.733159] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.733159] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.733159] env[61006]: ERROR nova.compute.manager raise self.value [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.733159] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.733159] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.733159] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.733735] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.733735] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.733735] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 651.733735] env[61006]: ERROR nova.compute.manager [ 651.733735] env[61006]: Traceback (most recent call last): [ 651.733735] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.733735] env[61006]: listener.cb(fileno) [ 651.733735] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.733735] env[61006]: result = function(*args, **kwargs) [ 651.733735] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.733735] env[61006]: return func(*args, **kwargs) [ 651.733735] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.733735] env[61006]: raise e [ 651.733735] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.733735] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 651.733735] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.733735] env[61006]: created_port_ids = self._update_ports_for_instance( [ 651.733735] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.733735] env[61006]: with excutils.save_and_reraise_exception(): [ 651.733735] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.733735] env[61006]: self.force_reraise() [ 651.733735] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.733735] env[61006]: raise self.value [ 651.733735] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.733735] env[61006]: updated_port = self._update_port( [ 651.733735] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.733735] env[61006]: _ensure_no_port_binding_failure(port) [ 651.733735] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.733735] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.734544] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 651.734544] env[61006]: Removing descriptor: 20 [ 651.739459] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04977bf3-9607-4575-8420-02ef3a611b60 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.756168] env[61006]: ERROR nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Traceback (most recent call last): [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] yield resources [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.driver.spawn(context, instance, image_meta, [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] vm_ref = self.build_virtual_machine(instance, [ 651.756168] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] for vif in network_info: [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return self._sync_wrapper(fn, *args, **kwargs) [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.wait() [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self[:] = self._gt.wait() [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return self._exit_event.wait() [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 651.756560] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] current.throw(*self._exc) [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] result = function(*args, **kwargs) [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return func(*args, **kwargs) [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise e [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] nwinfo = self.network_api.allocate_for_instance( [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] created_port_ids = self._update_ports_for_instance( [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] with excutils.save_and_reraise_exception(): [ 651.756913] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.force_reraise() [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise self.value [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] updated_port = self._update_port( [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] _ensure_no_port_binding_failure(port) [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise exception.PortBindingFailed(port_id=port['id']) [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 651.757267] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] [ 651.757267] env[61006]: INFO nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Terminating instance [ 651.762701] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.762878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquired lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.763061] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.038218] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494739} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.038531] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 652.038776] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.039061] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40ba2608-9f02-487a-8aa2-439ac1cfb238 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.045229] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 652.045229] env[61006]: value = "task-1336902" [ 652.045229] env[61006]: _type = "Task" [ 652.045229] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.052408] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.233735] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.234303] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.236838] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.008s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.238308] env[61006]: INFO nova.compute.claims [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.281210] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.334529] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.553165] env[61006]: DEBUG nova.compute.manager [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Received event network-changed-60aca473-1610-4bac-a857-d394c905c592 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.553328] env[61006]: DEBUG nova.compute.manager [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Refreshing instance network info cache due to event network-changed-60aca473-1610-4bac-a857-d394c905c592. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.553515] env[61006]: DEBUG oslo_concurrency.lockutils [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] Acquiring lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.559026] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065196} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.559745] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 652.560057] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8839b380-fd35-4611-93cd-c24c184a08f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.579270] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 652.579602] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-169c475c-2e43-43b9-82b4-b1da24f90a73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.598543] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 652.598543] env[61006]: value = "task-1336903" [ 652.598543] env[61006]: _type = "Task" [ 652.598543] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.607610] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.743508] env[61006]: DEBUG nova.compute.utils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.747476] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.747620] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.808385] env[61006]: DEBUG nova.policy [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdcaec59095a47e4acf3b832d7edf67f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b58e44ef36ba47f8ae64444ed6d0d964', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.838033] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Releasing lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.838268] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.838461] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.838764] env[61006]: DEBUG oslo_concurrency.lockutils [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] Acquired lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.838970] env[61006]: DEBUG nova.network.neutron [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Refreshing network info cache for port 60aca473-1610-4bac-a857-d394c905c592 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.840041] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73866044-7376-4a94-8d18-af71da0c5f0a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.850066] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aea2e5-7dd2-4064-83e7-92012e7dc240 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.873590] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5e4a8c02-be44-4139-b591-e1ed4ef52261 could not be found. [ 652.873822] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.874010] env[61006]: INFO nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.874331] env[61006]: DEBUG oslo.service.loopingcall [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.874558] env[61006]: DEBUG nova.compute.manager [-] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.874651] env[61006]: DEBUG nova.network.neutron [-] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.894977] env[61006]: DEBUG nova.network.neutron [-] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.109850] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336903, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.126724] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Successfully created port: 78908742-e120-4566-9bee-06b6cc15a8f3 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.248250] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.360933] env[61006]: DEBUG nova.network.neutron [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.401493] env[61006]: DEBUG nova.network.neutron [-] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.446796] env[61006]: DEBUG nova.network.neutron [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.612866] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336903, 'name': ReconfigVM_Task, 'duration_secs': 0.514267} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.616134] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c/47124964-8057-46ba-8c57-1a9f4471402c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 653.618033] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-690463e8-ab21-4e7c-959e-8cef747cc0fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.630517] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 653.630517] env[61006]: value = "task-1336904" [ 653.630517] env[61006]: _type = "Task" [ 653.630517] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.637831] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336904, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.706575] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e578aa6-3fd4-4038-9938-055ac68c26b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.714602] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852367f4-4927-4aa3-b12f-11fa0f1f4b3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.748350] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8612cee1-9c5c-4515-bfcd-aac209f7e5c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.758631] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a96f7df-ee2e-4ec9-bee6-a817810b6354 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.772741] env[61006]: DEBUG nova.compute.provider_tree [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.905926] env[61006]: INFO nova.compute.manager [-] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Took 1.03 seconds to deallocate network for instance. [ 653.907222] env[61006]: DEBUG nova.compute.claims [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.907566] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.949941] env[61006]: DEBUG oslo_concurrency.lockutils [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] Releasing lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.949941] env[61006]: DEBUG nova.compute.manager [req-c1f31d3c-c097-413f-b33a-b8e1b034a927 req-d5887abf-c139-436e-ac51-10d18e52f64d service nova] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Received event network-vif-deleted-60aca473-1610-4bac-a857-d394c905c592 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.133549] env[61006]: ERROR nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 654.133549] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.133549] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.133549] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.133549] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.133549] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.133549] env[61006]: ERROR nova.compute.manager raise self.value [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.133549] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.133549] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.133549] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.134057] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.134057] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.134057] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 654.134057] env[61006]: ERROR nova.compute.manager [ 654.134057] env[61006]: Traceback (most recent call last): [ 654.134057] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.134057] env[61006]: listener.cb(fileno) [ 654.134057] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.134057] env[61006]: result = function(*args, **kwargs) [ 654.134057] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.134057] env[61006]: return func(*args, **kwargs) [ 654.134057] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.134057] env[61006]: raise e [ 654.134057] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.134057] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 654.134057] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.134057] env[61006]: created_port_ids = self._update_ports_for_instance( [ 654.134057] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.134057] env[61006]: with excutils.save_and_reraise_exception(): [ 654.134057] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.134057] env[61006]: self.force_reraise() [ 654.134057] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.134057] env[61006]: raise self.value [ 654.134057] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.134057] env[61006]: updated_port = self._update_port( [ 654.134057] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.134057] env[61006]: _ensure_no_port_binding_failure(port) [ 654.134057] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.134057] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.134898] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 654.134898] env[61006]: Removing descriptor: 20 [ 654.137019] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336904, 'name': Rename_Task, 'duration_secs': 0.140338} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.137138] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 654.137372] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a44bd979-8c68-402d-9d7e-949186a990a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.143677] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 654.143677] env[61006]: value = "task-1336905" [ 654.143677] env[61006]: _type = "Task" [ 654.143677] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.151104] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.257599] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.275669] env[61006]: DEBUG nova.scheduler.client.report [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.286996] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.287270] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.287425] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.287605] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.287750] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.287898] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.288297] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.288484] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.288651] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.288808] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.288980] env[61006]: DEBUG nova.virt.hardware [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.289892] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ca5271-1d9a-47e8-8661-e22385dfc4f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.298466] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60e973b-957f-47e6-8314-437e54495ad3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.312630] env[61006]: ERROR nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] Traceback (most recent call last): [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] yield resources [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.driver.spawn(context, instance, image_meta, [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] vm_ref = self.build_virtual_machine(instance, [ 654.312630] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] for vif in network_info: [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return self._sync_wrapper(fn, *args, **kwargs) [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.wait() [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self[:] = self._gt.wait() [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return self._exit_event.wait() [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.313031] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] current.throw(*self._exc) [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] result = function(*args, **kwargs) [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return func(*args, **kwargs) [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise e [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] nwinfo = self.network_api.allocate_for_instance( [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] created_port_ids = self._update_ports_for_instance( [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] with excutils.save_and_reraise_exception(): [ 654.313407] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.force_reraise() [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise self.value [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] updated_port = self._update_port( [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] _ensure_no_port_binding_failure(port) [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise exception.PortBindingFailed(port_id=port['id']) [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 654.313771] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] [ 654.313771] env[61006]: INFO nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Terminating instance [ 654.315037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.315185] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquired lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.315351] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.580686] env[61006]: DEBUG nova.compute.manager [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Received event network-changed-78908742-e120-4566-9bee-06b6cc15a8f3 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.580936] env[61006]: DEBUG nova.compute.manager [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Refreshing instance network info cache due to event network-changed-78908742-e120-4566-9bee-06b6cc15a8f3. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 654.581148] env[61006]: DEBUG oslo_concurrency.lockutils [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] Acquiring lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.656307] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336905, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.783091] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.783811] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.786631] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.622s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.833715] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.891988] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.154542] env[61006]: DEBUG oslo_vmware.api [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336905, 'name': PowerOnVM_Task, 'duration_secs': 0.724439} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.154866] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 655.155059] env[61006]: DEBUG nova.compute.manager [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 655.155882] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db28903-ec49-4da8-b069-dd04d6cd3e38 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.292380] env[61006]: DEBUG nova.compute.utils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.297726] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.297726] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.346085] env[61006]: DEBUG nova.policy [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2867fd5b65d743e2a538f73855de1859', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28711bb635749fcbf46eb8b43327fec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.396428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Releasing lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.396993] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.397374] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.399979] env[61006]: DEBUG oslo_concurrency.lockutils [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] Acquired lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.400198] env[61006]: DEBUG nova.network.neutron [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Refreshing network info cache for port 78908742-e120-4566-9bee-06b6cc15a8f3 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.401330] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45158faf-f7e5-4644-90ca-c48cd9b58cb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.411534] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5dc117-5c78-46ef-82c8-aac3a030def9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.438194] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09646374-a3e3-4fc6-a227-950966144315 could not be found. [ 655.438420] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.438596] env[61006]: INFO nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.438838] env[61006]: DEBUG oslo.service.loopingcall [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.441372] env[61006]: DEBUG nova.compute.manager [-] [instance: 09646374-a3e3-4fc6-a227-950966144315] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.441467] env[61006]: DEBUG nova.network.neutron [-] [instance: 09646374-a3e3-4fc6-a227-950966144315] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.461503] env[61006]: DEBUG nova.network.neutron [-] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.674698] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.753140] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4949b8-7e3e-451d-81e8-7b5932cd213b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.761020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603ae0a1-3131-47ca-9a13-2bc9a341494d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.793081] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7735bc3b-986f-4983-9ae4-eb9cdbd99200 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.797492] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.803307] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2d2377-47d1-4b9c-896e-1a49dda6c73a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.823086] env[61006]: DEBUG nova.compute.provider_tree [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.882649] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Successfully created port: 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.930864] env[61006]: DEBUG nova.network.neutron [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.966525] env[61006]: DEBUG nova.network.neutron [-] [instance: 09646374-a3e3-4fc6-a227-950966144315] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.012691] env[61006]: DEBUG nova.network.neutron [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.088759] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "47124964-8057-46ba-8c57-1a9f4471402c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.089025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.089235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "47124964-8057-46ba-8c57-1a9f4471402c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.089417] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.089590] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.092257] env[61006]: INFO nova.compute.manager [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Terminating instance [ 656.094468] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "refresh_cache-47124964-8057-46ba-8c57-1a9f4471402c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.094625] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "refresh_cache-47124964-8057-46ba-8c57-1a9f4471402c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.094810] env[61006]: DEBUG nova.network.neutron [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.332036] env[61006]: DEBUG nova.scheduler.client.report [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.470931] env[61006]: INFO nova.compute.manager [-] [instance: 09646374-a3e3-4fc6-a227-950966144315] Took 1.03 seconds to deallocate network for instance. [ 656.471911] env[61006]: DEBUG nova.compute.claims [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.472285] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.515749] env[61006]: DEBUG oslo_concurrency.lockutils [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] Releasing lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.515749] env[61006]: DEBUG nova.compute.manager [req-aabe8f15-a949-416c-90f7-e6713e1ba75a req-e4bb3b7c-d6f6-41d1-987d-71354459e61c service nova] [instance: 09646374-a3e3-4fc6-a227-950966144315] Received event network-vif-deleted-78908742-e120-4566-9bee-06b6cc15a8f3 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.618198] env[61006]: DEBUG nova.network.neutron [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.702018] env[61006]: DEBUG nova.network.neutron [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.770817] env[61006]: DEBUG nova.compute.manager [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Received event network-changed-430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.771156] env[61006]: DEBUG nova.compute.manager [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Refreshing instance network info cache due to event network-changed-430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.771714] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] Acquiring lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.771714] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] Acquired lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.771714] env[61006]: DEBUG nova.network.neutron [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Refreshing network info cache for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 656.812124] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.834733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.835669] env[61006]: ERROR nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Traceback (most recent call last): [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.driver.spawn(context, instance, image_meta, [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] vm_ref = self.build_virtual_machine(instance, [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.835669] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] for vif in network_info: [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self._sync_wrapper(fn, *args, **kwargs) [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.wait() [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self[:] = self._gt.wait() [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self._exit_event.wait() [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] result = hub.switch() [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.836050] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return self.greenlet.switch() [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] result = function(*args, **kwargs) [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] return func(*args, **kwargs) [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise e [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] nwinfo = self.network_api.allocate_for_instance( [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] created_port_ids = self._update_ports_for_instance( [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] with excutils.save_and_reraise_exception(): [ 656.836486] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] self.force_reraise() [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise self.value [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] updated_port = self._update_port( [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] _ensure_no_port_binding_failure(port) [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] raise exception.PortBindingFailed(port_id=port['id']) [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] nova.exception.PortBindingFailed: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. [ 656.836899] env[61006]: ERROR nova.compute.manager [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] [ 656.837287] env[61006]: DEBUG nova.compute.utils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.839564] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.839796] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.839971] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.840173] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.840316] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.840459] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.840656] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.840833] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.841055] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.841310] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.841393] env[61006]: DEBUG nova.virt.hardware [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.841681] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.962s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.844705] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f31a6a2-a80d-4256-9b97-a138058dc33b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.847932] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Build of instance 1fdae918-be3c-441d-8fee-ff5a2a32373c was re-scheduled: Binding failed for port f7794882-a3ef-42b9-87ea-1ec4119afa11, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.848445] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.848620] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.848764] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.848917] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.857145] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0e25a0-4a69-4117-a7ac-47f9cc2a6c8e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.910016] env[61006]: ERROR nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 656.910016] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.910016] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.910016] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.910016] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.910016] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.910016] env[61006]: ERROR nova.compute.manager raise self.value [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.910016] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.910016] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.910016] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.910486] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.910486] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.910486] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 656.910486] env[61006]: ERROR nova.compute.manager [ 656.910486] env[61006]: Traceback (most recent call last): [ 656.910486] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.910486] env[61006]: listener.cb(fileno) [ 656.910486] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.910486] env[61006]: result = function(*args, **kwargs) [ 656.910486] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.910486] env[61006]: return func(*args, **kwargs) [ 656.910486] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.910486] env[61006]: raise e [ 656.910486] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.910486] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 656.910486] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.910486] env[61006]: created_port_ids = self._update_ports_for_instance( [ 656.910486] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.910486] env[61006]: with excutils.save_and_reraise_exception(): [ 656.910486] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.910486] env[61006]: self.force_reraise() [ 656.910486] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.910486] env[61006]: raise self.value [ 656.910486] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.910486] env[61006]: updated_port = self._update_port( [ 656.910486] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.910486] env[61006]: _ensure_no_port_binding_failure(port) [ 656.910486] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.910486] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.911191] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 656.911191] env[61006]: Removing descriptor: 20 [ 656.911191] env[61006]: ERROR nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Traceback (most recent call last): [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] yield resources [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.driver.spawn(context, instance, image_meta, [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.911191] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] vm_ref = self.build_virtual_machine(instance, [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] for vif in network_info: [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self._sync_wrapper(fn, *args, **kwargs) [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.wait() [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self[:] = self._gt.wait() [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self._exit_event.wait() [ 656.911475] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] result = hub.switch() [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self.greenlet.switch() [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] result = function(*args, **kwargs) [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return func(*args, **kwargs) [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise e [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] nwinfo = self.network_api.allocate_for_instance( [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.911768] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] created_port_ids = self._update_ports_for_instance( [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] with excutils.save_and_reraise_exception(): [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.force_reraise() [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise self.value [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] updated_port = self._update_port( [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] _ensure_no_port_binding_failure(port) [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.912077] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise exception.PortBindingFailed(port_id=port['id']) [ 656.912367] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 656.912367] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] [ 656.912367] env[61006]: INFO nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Terminating instance [ 656.917049] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.204876] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "refresh_cache-47124964-8057-46ba-8c57-1a9f4471402c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.205404] env[61006]: DEBUG nova.compute.manager [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.205637] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.206570] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5155bad8-627a-4550-8fa9-8c08fb14c1d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.214930] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 657.215180] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd5ce25b-9652-4700-8c48-66f5a92c90d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.221067] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 657.221067] env[61006]: value = "task-1336906" [ 657.221067] env[61006]: _type = "Task" [ 657.221067] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.228928] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.294462] env[61006]: DEBUG nova.network.neutron [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.375212] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.404386] env[61006]: DEBUG nova.network.neutron [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.477823] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.731678] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336906, 'name': PowerOffVM_Task, 'duration_secs': 0.188273} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.732007] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 657.732268] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 657.732522] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c087fa1-d874-459d-8cfb-7718506201ac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.755989] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 657.756409] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 657.756409] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleting the datastore file [datastore2] 47124964-8057-46ba-8c57-1a9f4471402c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.756708] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26fc4192-493c-480d-a305-5bcd652071f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.762555] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 657.762555] env[61006]: value = "task-1336908" [ 657.762555] env[61006]: _type = "Task" [ 657.762555] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.770211] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.881369] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6642ffcb-cb01-4e38-a27c-bf4e4c938a17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 657.907079] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce14666c-b617-4ac0-8d2d-35b58eb14fce req-9670d63e-b24d-4491-84ef-7c32de3bb62e service nova] Releasing lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.907475] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.907657] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.983378] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-1fdae918-be3c-441d-8fee-ff5a2a32373c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.983625] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.983806] env[61006]: DEBUG nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.983972] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.004220] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.272273] env[61006]: DEBUG oslo_vmware.api [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096917} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.273026] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 658.273026] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 658.273026] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 658.273549] env[61006]: INFO nova.compute.manager [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Took 1.07 seconds to destroy the instance on the hypervisor. [ 658.273549] env[61006]: DEBUG oslo.service.loopingcall [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.273549] env[61006]: DEBUG nova.compute.manager [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.273686] env[61006]: DEBUG nova.network.neutron [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.292429] env[61006]: DEBUG nova.network.neutron [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.384205] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1fdae918-be3c-441d-8fee-ff5a2a32373c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.386018] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e2ef686c-b517-4121-88f5-01cad6b960b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386018] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 47124964-8057-46ba-8c57-1a9f4471402c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386018] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance aca01468-634e-4013-8a96-c7139f007c60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386018] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e42de415-1d9d-4931-a5b7-97112af42fd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386234] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance cf4e241c-a4e3-438c-8059-c7f7cb41cf45 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386234] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 5e4a8c02-be44-4139-b591-e1ed4ef52261 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386234] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 09646374-a3e3-4fc6-a227-950966144315 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.386234] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 8102112b-ee4d-4c7e-9b67-9b6a9c032f56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 658.423569] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.501822] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.506384] env[61006]: DEBUG nova.network.neutron [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.795289] env[61006]: DEBUG nova.network.neutron [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.804659] env[61006]: DEBUG nova.compute.manager [req-e3b65dc0-1f5e-453e-909e-0e008370c5e5 req-923fe7f1-2205-4564-87d1-4d72ab063a19 service nova] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Received event network-vif-deleted-430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.888245] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 5a2ddd3d-c939-4f20-a40d-108e805d64c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.006081] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Releasing lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.006197] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.006393] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 659.006684] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-136d9453-f6b8-490f-b145-1348a3dfbb03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.008990] env[61006]: INFO nova.compute.manager [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 1fdae918-be3c-441d-8fee-ff5a2a32373c] Took 1.02 seconds to deallocate network for instance. [ 659.018199] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f292c53-138e-418a-bc50-978ff2a31e59 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.043311] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8102112b-ee4d-4c7e-9b67-9b6a9c032f56 could not be found. [ 659.043556] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.043768] env[61006]: INFO nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.043972] env[61006]: DEBUG oslo.service.loopingcall [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.044370] env[61006]: DEBUG nova.compute.manager [-] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.044469] env[61006]: DEBUG nova.network.neutron [-] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.058227] env[61006]: DEBUG nova.network.neutron [-] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.298045] env[61006]: INFO nova.compute.manager [-] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Took 1.02 seconds to deallocate network for instance. [ 659.390813] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1542f4da-16bb-47c3-b11a-a9ac332ac917 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.560797] env[61006]: DEBUG nova.network.neutron [-] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.805609] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.893560] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance c0b97c94-a055-4ce3-b961-1cdb6a5955a3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.037852] env[61006]: INFO nova.scheduler.client.report [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted allocations for instance 1fdae918-be3c-441d-8fee-ff5a2a32373c [ 660.063182] env[61006]: INFO nova.compute.manager [-] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Took 1.02 seconds to deallocate network for instance. [ 660.065202] env[61006]: DEBUG nova.compute.claims [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.065377] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.397430] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance f7325d28-ed07-4dab-b440-2daf766882de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.545384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-79335ae0-5660-484e-82b0-1eeece3032e8 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "1fdae918-be3c-441d-8fee-ff5a2a32373c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.744s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.900151] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 5525a859-eb09-45a6-80d6-8a466976a7bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.048098] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.403541] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d6892ecb-0fb8-44da-9362-8806c429626b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.455029] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.455216] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.573317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.906385] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance ea2a73fa-3ece-4c88-b117-3b9843a4f224 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.409772] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 34f17493-0d3d-4dfd-a2ac-6adc9b388f95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.912414] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 2217133d-52fb-45c4-bb84-63dde0546747 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.416098] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0211d621-ba0e-480b-8abd-58df829a1e39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.918117] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 232272f0-1bf5-436e-ae24-5efa391eef57 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.421405] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6c156407-4ac6-4e34-812d-d719f1cc33d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.925585] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance fd415afa-1e5f-42ea-b288-4432f15a5503 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.427734] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4a38192e-4fb8-4767-b1d9-3d8b59a76925 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.930756] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance b8fb9290-755f-40b8-af3d-c17407d0c846 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.433528] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.936396] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0ad1f43f-9e98-43b7-8076-20824ccc18ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.440219] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e4a5a57e-f078-48ec-afe5-204c75139c15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.944544] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d07742fb-259a-4173-84df-c7c40838ba2c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.447542] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1ebc21ae-d46f-486b-952a-61324aaf353f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.951087] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance dac63cd3-1788-4e45-b58a-9cd22038dd30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.454470] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6ee89cb9-ef41-4c85-98d2-5b0190568efe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.957169] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance cca6374c-09a4-4145-a116-c49c5a8330c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.460327] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6e99894e-81b6-4a07-9ec7-caa16272b3ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.460622] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 670.460767] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 670.841799] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efe967f-144e-48db-910f-94540be1c1c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.849343] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e529ab-941c-4d3c-bb6b-ff58c1b6e70c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.878900] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e0542d-0f11-46f3-9aaf-8bcb935b9a43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.886226] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9957e9-cd64-4c66-aa8b-f24cdbd718c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.898834] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.401638] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.906787] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 671.907083] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.065s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.907372] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.016s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.796306] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dce6df2-c404-408a-bd13-ff905486bda6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.803717] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddb36b6-55d3-4551-83c4-c96e832803ae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.831886] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68591654-6af2-48af-bfaf-f8a26748ad84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.839056] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5f0a1d-2ff1-4ca7-b3b7-be7366b13b45 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.852729] env[61006]: DEBUG nova.compute.provider_tree [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.355795] env[61006]: DEBUG nova.scheduler.client.report [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.861825] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.862574] env[61006]: ERROR nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Traceback (most recent call last): [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.driver.spawn(context, instance, image_meta, [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] vm_ref = self.build_virtual_machine(instance, [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.862574] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] for vif in network_info: [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self._sync_wrapper(fn, *args, **kwargs) [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.wait() [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self[:] = self._gt.wait() [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self._exit_event.wait() [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] result = hub.switch() [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.862875] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return self.greenlet.switch() [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] result = function(*args, **kwargs) [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] return func(*args, **kwargs) [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise e [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] nwinfo = self.network_api.allocate_for_instance( [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] created_port_ids = self._update_ports_for_instance( [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] with excutils.save_and_reraise_exception(): [ 673.863205] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] self.force_reraise() [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise self.value [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] updated_port = self._update_port( [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] _ensure_no_port_binding_failure(port) [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] raise exception.PortBindingFailed(port_id=port['id']) [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] nova.exception.PortBindingFailed: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. [ 673.863503] env[61006]: ERROR nova.compute.manager [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] [ 673.863763] env[61006]: DEBUG nova.compute.utils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.864672] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.879s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.867651] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Build of instance e2ef686c-b517-4121-88f5-01cad6b960b9 was re-scheduled: Binding failed for port bb89e6fb-b08d-46f8-a99c-ef8b5eb14f2d, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.867813] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.868089] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.868243] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.868398] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.388677] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.473694] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.748336] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720e07fe-7adb-489d-9312-c7968f75d510 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.756192] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef851e29-e0ee-42c5-bdd7-617792248a1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.785489] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e76db08-bb58-492f-93df-b0a8434dd9a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.792757] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e93785-6f0d-4ed7-bc5c-e55deece8988 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.806902] env[61006]: DEBUG nova.compute.provider_tree [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.978391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "refresh_cache-e2ef686c-b517-4121-88f5-01cad6b960b9" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.978675] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.978822] env[61006]: DEBUG nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.978986] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.002539] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.310294] env[61006]: DEBUG nova.scheduler.client.report [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.505490] env[61006]: DEBUG nova.network.neutron [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.815770] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.816265] env[61006]: ERROR nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] Traceback (most recent call last): [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.driver.spawn(context, instance, image_meta, [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] vm_ref = self.build_virtual_machine(instance, [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.816265] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] for vif in network_info: [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return self._sync_wrapper(fn, *args, **kwargs) [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.wait() [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self[:] = self._gt.wait() [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return self._exit_event.wait() [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] current.throw(*self._exc) [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.816562] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] result = function(*args, **kwargs) [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] return func(*args, **kwargs) [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise e [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] nwinfo = self.network_api.allocate_for_instance( [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] created_port_ids = self._update_ports_for_instance( [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] with excutils.save_and_reraise_exception(): [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] self.force_reraise() [ 675.816868] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise self.value [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] updated_port = self._update_port( [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] _ensure_no_port_binding_failure(port) [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] raise exception.PortBindingFailed(port_id=port['id']) [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] nova.exception.PortBindingFailed: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. [ 675.817336] env[61006]: ERROR nova.compute.manager [instance: aca01468-634e-4013-8a96-c7139f007c60] [ 675.817336] env[61006]: DEBUG nova.compute.utils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.818316] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.160s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.819879] env[61006]: INFO nova.compute.claims [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.822826] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Build of instance aca01468-634e-4013-8a96-c7139f007c60 was re-scheduled: Binding failed for port d6fbd2b9-e7b0-43d2-8e87-a7e61e83ec40, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.823180] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.823406] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquiring lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.823550] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Acquired lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.823703] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.008120] env[61006]: INFO nova.compute.manager [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: e2ef686c-b517-4121-88f5-01cad6b960b9] Took 1.03 seconds to deallocate network for instance. [ 676.345789] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.425816] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.928227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Releasing lock "refresh_cache-aca01468-634e-4013-8a96-c7139f007c60" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.928501] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.928684] env[61006]: DEBUG nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.928850] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.951517] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.039251] env[61006]: INFO nova.scheduler.client.report [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Deleted allocations for instance e2ef686c-b517-4121-88f5-01cad6b960b9 [ 677.251526] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478f3d33-a9ce-4510-888d-f7b228fbeb6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.259029] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c36bb6-eb61-4a3b-809f-f57d085e3db9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.289294] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39936e49-4aec-4dba-b09b-21493d8ed061 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.296340] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bd2de4-7438-4de4-88de-b9993fdfeb6a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.308919] env[61006]: DEBUG nova.compute.provider_tree [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.456998] env[61006]: DEBUG nova.network.neutron [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.550695] env[61006]: DEBUG oslo_concurrency.lockutils [None req-289d4aeb-aaaa-43aa-9765-9ad9006b86a3 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "e2ef686c-b517-4121-88f5-01cad6b960b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.732s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.812559] env[61006]: DEBUG nova.scheduler.client.report [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.959378] env[61006]: INFO nova.compute.manager [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] [instance: aca01468-634e-4013-8a96-c7139f007c60] Took 1.03 seconds to deallocate network for instance. [ 678.052991] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.318238] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.318602] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.321323] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.629s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.322702] env[61006]: INFO nova.compute.claims [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.575831] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.829230] env[61006]: DEBUG nova.compute.utils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.830965] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.831167] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 678.896255] env[61006]: DEBUG nova.policy [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab8f1f51e564fae9772a278317b8a9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '366353922be146f38fdfedbc9e949259', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.987489] env[61006]: INFO nova.scheduler.client.report [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Deleted allocations for instance aca01468-634e-4013-8a96-c7139f007c60 [ 679.284319] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Successfully created port: 2d065202-2687-4ff8-87c4-adbb2b7099d8 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.335271] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.499151] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65b71a09-7c19-4e8d-8bc3-9328e821ade6 tempest-FloatingIPsAssociationTestJSON-518090047 tempest-FloatingIPsAssociationTestJSON-518090047-project-member] Lock "aca01468-634e-4013-8a96-c7139f007c60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.007s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.793967] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc2c713-d56e-4260-9376-881356a6ab9e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.807084] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a5ada7-9d4c-4bdf-9d1f-d240f0e57b81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.838076] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8622074d-3e1e-4a28-9970-4e3f13a720f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.841957] env[61006]: INFO nova.virt.block_device [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Booting with volume d0c3d4d2-044a-4d94-afb9-56c898fee1bf at /dev/sda [ 679.863242] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b05c66a-aec8-4d8c-91ae-38d12a2e14a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.877436] env[61006]: DEBUG nova.compute.provider_tree [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.893895] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e087a6f3-dfe7-4676-8275-3ea995e22b09 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.902562] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fdec65-0993-4c2a-88c2-ce4058f67d3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.924601] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68f7361c-be8c-4a9d-be35-10524d2e8cb1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.932765] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca7ec46-60dc-4e1f-a09c-3b0a3f44398e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.954081] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e22eea6-8e54-44c6-9bec-b0e8622a0d52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.960217] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20daf8dd-1739-46fc-af3e-55c7b961a99e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.974069] env[61006]: DEBUG nova.virt.block_device [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating existing volume attachment record: 0f66e500-2063-47af-8ecc-27d692cd833c {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 680.001656] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.173857] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "88e2bdc9-ab73-4e23-94b5-a45046835144" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.174104] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.303013] env[61006]: DEBUG nova.compute.manager [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Received event network-changed-2d065202-2687-4ff8-87c4-adbb2b7099d8 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.303264] env[61006]: DEBUG nova.compute.manager [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Refreshing instance network info cache due to event network-changed-2d065202-2687-4ff8-87c4-adbb2b7099d8. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.303419] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] Acquiring lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.303558] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] Acquired lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.303710] env[61006]: DEBUG nova.network.neutron [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Refreshing network info cache for port 2d065202-2687-4ff8-87c4-adbb2b7099d8 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.381204] env[61006]: DEBUG nova.scheduler.client.report [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.532900] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.608567] env[61006]: ERROR nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 680.608567] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.608567] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.608567] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.608567] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.608567] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.608567] env[61006]: ERROR nova.compute.manager raise self.value [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.608567] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.608567] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.608567] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.609127] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.609127] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.609127] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 680.609127] env[61006]: ERROR nova.compute.manager [ 680.609127] env[61006]: Traceback (most recent call last): [ 680.609127] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.609127] env[61006]: listener.cb(fileno) [ 680.609127] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.609127] env[61006]: result = function(*args, **kwargs) [ 680.609127] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.609127] env[61006]: return func(*args, **kwargs) [ 680.609127] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.609127] env[61006]: raise e [ 680.609127] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.609127] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 680.609127] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.609127] env[61006]: created_port_ids = self._update_ports_for_instance( [ 680.609127] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.609127] env[61006]: with excutils.save_and_reraise_exception(): [ 680.609127] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.609127] env[61006]: self.force_reraise() [ 680.609127] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.609127] env[61006]: raise self.value [ 680.609127] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.609127] env[61006]: updated_port = self._update_port( [ 680.609127] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.609127] env[61006]: _ensure_no_port_binding_failure(port) [ 680.609127] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.609127] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.609949] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 680.609949] env[61006]: Removing descriptor: 17 [ 680.827485] env[61006]: DEBUG nova.network.neutron [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.886396] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.565s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.886907] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.889516] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.165s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.987897] env[61006]: DEBUG nova.network.neutron [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.398368] env[61006]: DEBUG nova.compute.utils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.410078] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 681.410927] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 681.490775] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf2f993e-fadf-43de-8d42-f92de5005158 req-58df36ea-335f-458e-afd7-31918533fbc3 service nova] Releasing lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.494880] env[61006]: DEBUG nova.policy [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ec2836730b8546c296aa9e3b0e3bafaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b6ae2433d39843348a88ff978af4a2bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.879845] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Successfully created port: 6f73dae8-091b-49cc-902a-3aa360a5d6ca {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.889671] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46163248-9822-4d1e-bca0-8347983d1b28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.896445] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f172b5cf-dd7c-440f-b118-1b53e6696b9e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.933605] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.936696] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfc2008-275a-4993-a9f0-b8e77993ed8e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.944407] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946dca34-3b82-470e-a52e-47f07b491692 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.957714] env[61006]: DEBUG nova.compute.provider_tree [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.101027] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.101027] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.101027] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.101027] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.101260] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.101260] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.101260] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.101595] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.101866] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.102149] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.102459] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.102735] env[61006]: DEBUG nova.virt.hardware [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.103748] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25bb554-f005-4809-8724-773aa3ce257e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.112684] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc4dae9-c09a-4137-9732-11f9fc2488c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.128386] env[61006]: ERROR nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Traceback (most recent call last): [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] yield resources [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.driver.spawn(context, instance, image_meta, [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] vm_ref = self.build_virtual_machine(instance, [ 682.128386] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] for vif in network_info: [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return self._sync_wrapper(fn, *args, **kwargs) [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.wait() [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self[:] = self._gt.wait() [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return self._exit_event.wait() [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.129189] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] current.throw(*self._exc) [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] result = function(*args, **kwargs) [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return func(*args, **kwargs) [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise e [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] nwinfo = self.network_api.allocate_for_instance( [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] created_port_ids = self._update_ports_for_instance( [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] with excutils.save_and_reraise_exception(): [ 682.129586] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.force_reraise() [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise self.value [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] updated_port = self._update_port( [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] _ensure_no_port_binding_failure(port) [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise exception.PortBindingFailed(port_id=port['id']) [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 682.129965] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] [ 682.129965] env[61006]: INFO nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Terminating instance [ 682.132870] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquiring lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.133047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquired lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.133391] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.333014] env[61006]: DEBUG nova.compute.manager [req-21e9aea7-00a9-42a6-a2de-9fcb244a7f62 req-d612dc5e-3688-4fea-82d5-73967feca777 service nova] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Received event network-vif-deleted-2d065202-2687-4ff8-87c4-adbb2b7099d8 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.460856] env[61006]: DEBUG nova.scheduler.client.report [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.655323] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.796885] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.953171] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.965713] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.966672] env[61006]: ERROR nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Traceback (most recent call last): [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.driver.spawn(context, instance, image_meta, [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] vm_ref = self.build_virtual_machine(instance, [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.966672] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] for vif in network_info: [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return self._sync_wrapper(fn, *args, **kwargs) [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.wait() [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self[:] = self._gt.wait() [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return self._exit_event.wait() [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] current.throw(*self._exc) [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.967026] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] result = function(*args, **kwargs) [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] return func(*args, **kwargs) [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise e [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] nwinfo = self.network_api.allocate_for_instance( [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] created_port_ids = self._update_ports_for_instance( [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] with excutils.save_and_reraise_exception(): [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] self.force_reraise() [ 682.968059] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise self.value [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] updated_port = self._update_port( [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] _ensure_no_port_binding_failure(port) [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] raise exception.PortBindingFailed(port_id=port['id']) [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] nova.exception.PortBindingFailed: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. [ 682.968549] env[61006]: ERROR nova.compute.manager [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] [ 682.968549] env[61006]: DEBUG nova.compute.utils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.969053] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Build of instance e42de415-1d9d-4931-a5b7-97112af42fd1 was re-scheduled: Binding failed for port a122f146-2428-4908-938f-268937fa692f, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.969521] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.969810] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquiring lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.970082] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Acquired lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.970341] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.971556] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.623s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.982837] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.983111] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.984774] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.984774] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.984774] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.984915] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.985192] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.985396] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.985626] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.985843] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.986344] env[61006]: DEBUG nova.virt.hardware [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.987093] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1732dfc6-3179-4781-9beb-783e6118016f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.998424] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8e7555-11b1-484e-a44e-fcf5dca3fa31 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.018151] env[61006]: ERROR nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 683.018151] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.018151] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.018151] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.018151] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.018151] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.018151] env[61006]: ERROR nova.compute.manager raise self.value [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.018151] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.018151] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.018151] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.018595] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.018595] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.018595] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 683.018595] env[61006]: ERROR nova.compute.manager [ 683.018595] env[61006]: Traceback (most recent call last): [ 683.018595] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.018595] env[61006]: listener.cb(fileno) [ 683.018595] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.018595] env[61006]: result = function(*args, **kwargs) [ 683.018595] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.018595] env[61006]: return func(*args, **kwargs) [ 683.018595] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.018595] env[61006]: raise e [ 683.018595] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.018595] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 683.018595] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.018595] env[61006]: created_port_ids = self._update_ports_for_instance( [ 683.018595] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.018595] env[61006]: with excutils.save_and_reraise_exception(): [ 683.018595] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.018595] env[61006]: self.force_reraise() [ 683.018595] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.018595] env[61006]: raise self.value [ 683.018595] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.018595] env[61006]: updated_port = self._update_port( [ 683.018595] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.018595] env[61006]: _ensure_no_port_binding_failure(port) [ 683.018595] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.018595] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.019412] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 683.019412] env[61006]: Removing descriptor: 17 [ 683.019412] env[61006]: ERROR nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Traceback (most recent call last): [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] yield resources [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.driver.spawn(context, instance, image_meta, [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.019412] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] vm_ref = self.build_virtual_machine(instance, [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] for vif in network_info: [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self._sync_wrapper(fn, *args, **kwargs) [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.wait() [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self[:] = self._gt.wait() [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self._exit_event.wait() [ 683.019709] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] result = hub.switch() [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self.greenlet.switch() [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] result = function(*args, **kwargs) [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return func(*args, **kwargs) [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise e [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] nwinfo = self.network_api.allocate_for_instance( [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.020089] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] created_port_ids = self._update_ports_for_instance( [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] with excutils.save_and_reraise_exception(): [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.force_reraise() [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise self.value [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] updated_port = self._update_port( [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] _ensure_no_port_binding_failure(port) [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.020415] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise exception.PortBindingFailed(port_id=port['id']) [ 683.020701] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 683.020701] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] [ 683.020701] env[61006]: INFO nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Terminating instance [ 683.021526] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquiring lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.021760] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquired lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.021974] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.301768] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Releasing lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.302392] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.302746] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e72f7c7c-342e-42fc-abde-dcc780932495 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.310992] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce0e10a-f133-44ef-8825-0dfd5fbf3353 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.333900] env[61006]: WARNING nova.virt.vmwareapi.driver [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5a2ddd3d-c939-4f20-a40d-108e805d64c4 could not be found. [ 683.333900] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.333900] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42eed157-af24-4592-af53-4d093fd3ec85 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.341568] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb337fa-1e59-4b7a-95e6-ed51ae02f4c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.363739] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5a2ddd3d-c939-4f20-a40d-108e805d64c4 could not be found. [ 683.363966] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.364158] env[61006]: INFO nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Took 0.06 seconds to destroy the instance on the hypervisor. [ 683.364399] env[61006]: DEBUG oslo.service.loopingcall [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.364699] env[61006]: DEBUG nova.compute.manager [-] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.364699] env[61006]: DEBUG nova.network.neutron [-] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.378598] env[61006]: DEBUG nova.network.neutron [-] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.513159] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.553932] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.657690] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.673819] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.880712] env[61006]: DEBUG nova.network.neutron [-] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.897730] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6795a2e3-7c73-41f9-8e38-e67cb6abb5f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.906214] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab60b457-acc5-408d-ac78-2c463655a4d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.939735] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0196ae-87b8-441e-8ee9-fe850ab97f12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.947017] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2957f5fe-0d91-452e-8419-ce72905b90f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.960904] env[61006]: DEBUG nova.compute.provider_tree [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.160474] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Releasing lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.160886] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 684.161083] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.161407] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1cde6ea-4900-4155-8998-ba4cc1d81f66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.171424] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e4cc89-b0a9-41c1-9998-d7b0b20625fa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.181609] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Releasing lock "refresh_cache-e42de415-1d9d-4931-a5b7-97112af42fd1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.181764] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.181831] env[61006]: DEBUG nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.182147] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.195831] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1542f4da-16bb-47c3-b11a-a9ac332ac917 could not be found. [ 684.196115] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.196249] env[61006]: INFO nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Took 0.04 seconds to destroy the instance on the hypervisor. [ 684.196492] env[61006]: DEBUG oslo.service.loopingcall [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.196703] env[61006]: DEBUG nova.compute.manager [-] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.196794] env[61006]: DEBUG nova.network.neutron [-] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.198967] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.221750] env[61006]: DEBUG nova.network.neutron [-] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.369502] env[61006]: DEBUG nova.compute.manager [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Received event network-changed-6f73dae8-091b-49cc-902a-3aa360a5d6ca {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.369688] env[61006]: DEBUG nova.compute.manager [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Refreshing instance network info cache due to event network-changed-6f73dae8-091b-49cc-902a-3aa360a5d6ca. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.369891] env[61006]: DEBUG oslo_concurrency.lockutils [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] Acquiring lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.370033] env[61006]: DEBUG oslo_concurrency.lockutils [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] Acquired lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.370284] env[61006]: DEBUG nova.network.neutron [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Refreshing network info cache for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.383437] env[61006]: INFO nova.compute.manager [-] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Took 1.02 seconds to deallocate network for instance. [ 684.464188] env[61006]: DEBUG nova.scheduler.client.report [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.702319] env[61006]: DEBUG nova.network.neutron [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.723670] env[61006]: DEBUG nova.network.neutron [-] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.890724] env[61006]: DEBUG nova.network.neutron [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.948276] env[61006]: DEBUG nova.network.neutron [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.950860] env[61006]: INFO nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Took 0.57 seconds to detach 1 volumes for instance. [ 684.953280] env[61006]: DEBUG nova.compute.claims [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.953472] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.970169] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.970777] env[61006]: ERROR nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Traceback (most recent call last): [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.driver.spawn(context, instance, image_meta, [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] vm_ref = self.build_virtual_machine(instance, [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.970777] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] for vif in network_info: [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return self._sync_wrapper(fn, *args, **kwargs) [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.wait() [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self[:] = self._gt.wait() [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return self._exit_event.wait() [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] current.throw(*self._exc) [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.971159] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] result = function(*args, **kwargs) [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] return func(*args, **kwargs) [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise e [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] nwinfo = self.network_api.allocate_for_instance( [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] created_port_ids = self._update_ports_for_instance( [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] with excutils.save_and_reraise_exception(): [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] self.force_reraise() [ 684.971518] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise self.value [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] updated_port = self._update_port( [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] _ensure_no_port_binding_failure(port) [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] raise exception.PortBindingFailed(port_id=port['id']) [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] nova.exception.PortBindingFailed: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. [ 684.971850] env[61006]: ERROR nova.compute.manager [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] [ 684.971850] env[61006]: DEBUG nova.compute.utils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.972523] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.065s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.977412] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Build of instance cf4e241c-a4e3-438c-8059-c7f7cb41cf45 was re-scheduled: Binding failed for port 72a7c63c-3585-4b4a-bf51-f6ebb89a232f, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.978149] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.978149] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquiring lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.978280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Acquired lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.979413] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.205136] env[61006]: INFO nova.compute.manager [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] [instance: e42de415-1d9d-4931-a5b7-97112af42fd1] Took 1.02 seconds to deallocate network for instance. [ 685.226569] env[61006]: INFO nova.compute.manager [-] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Took 1.03 seconds to deallocate network for instance. [ 685.229897] env[61006]: DEBUG nova.compute.claims [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 685.229897] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.453703] env[61006]: DEBUG oslo_concurrency.lockutils [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] Releasing lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.453989] env[61006]: DEBUG nova.compute.manager [req-2310dea2-330a-4967-a9e3-5bdb2e40ef5f req-735c91d4-8df5-41da-8372-e76120dd5981 service nova] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Received event network-vif-deleted-6f73dae8-091b-49cc-902a-3aa360a5d6ca {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.509403] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.595519] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.886444] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.886716] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.965867] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c98049-0d88-45e7-9b05-1b62fda747bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.973585] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf0ca33-cba7-458e-9076-15051e9bb4da {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.002104] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dece7da-a548-4903-8ddb-f6e6b5bf992f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.010015] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4767d5f8-7f77-4c82-b325-f60949ca6887 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.023105] env[61006]: DEBUG nova.compute.provider_tree [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.098827] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Releasing lock "refresh_cache-cf4e241c-a4e3-438c-8059-c7f7cb41cf45" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.099077] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.099267] env[61006]: DEBUG nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.099403] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.129283] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.245888] env[61006]: INFO nova.scheduler.client.report [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Deleted allocations for instance e42de415-1d9d-4931-a5b7-97112af42fd1 [ 686.526218] env[61006]: DEBUG nova.scheduler.client.report [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.632927] env[61006]: DEBUG nova.network.neutron [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.758712] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9a465d5-ac92-411b-9c3c-cb41a5c5109c tempest-ServerAddressesTestJSON-1523738001 tempest-ServerAddressesTestJSON-1523738001-project-member] Lock "e42de415-1d9d-4931-a5b7-97112af42fd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.654s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.031383] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.059s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.032022] env[61006]: ERROR nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Traceback (most recent call last): [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.driver.spawn(context, instance, image_meta, [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] vm_ref = self.build_virtual_machine(instance, [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.032022] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] for vif in network_info: [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return self._sync_wrapper(fn, *args, **kwargs) [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.wait() [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self[:] = self._gt.wait() [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return self._exit_event.wait() [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] current.throw(*self._exc) [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.032330] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] result = function(*args, **kwargs) [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] return func(*args, **kwargs) [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise e [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] nwinfo = self.network_api.allocate_for_instance( [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] created_port_ids = self._update_ports_for_instance( [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] with excutils.save_and_reraise_exception(): [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] self.force_reraise() [ 687.032637] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise self.value [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] updated_port = self._update_port( [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] _ensure_no_port_binding_failure(port) [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] raise exception.PortBindingFailed(port_id=port['id']) [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] nova.exception.PortBindingFailed: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. [ 687.032950] env[61006]: ERROR nova.compute.manager [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] [ 687.032950] env[61006]: DEBUG nova.compute.utils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.034363] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Build of instance 5e4a8c02-be44-4139-b591-e1ed4ef52261 was re-scheduled: Binding failed for port 60aca473-1610-4bac-a857-d394c905c592, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.034752] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.034970] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.035127] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquired lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.035284] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.036523] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.362s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.036697] env[61006]: DEBUG nova.objects.instance [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61006) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 687.136050] env[61006]: INFO nova.compute.manager [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] [instance: cf4e241c-a4e3-438c-8059-c7f7cb41cf45] Took 1.04 seconds to deallocate network for instance. [ 687.260809] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.564563] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.637426] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.784220] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.049750] env[61006]: DEBUG oslo_concurrency.lockutils [None req-46e80ded-2265-4412-9cb9-ef4d0fe34beb tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.050965] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.579s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.143665] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Releasing lock "refresh_cache-5e4a8c02-be44-4139-b591-e1ed4ef52261" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.144401] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.144401] env[61006]: DEBUG nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.144401] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.168054] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.171224] env[61006]: INFO nova.scheduler.client.report [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Deleted allocations for instance cf4e241c-a4e3-438c-8059-c7f7cb41cf45 [ 688.670215] env[61006]: DEBUG nova.network.neutron [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.679798] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40c7acf4-8b8b-40a2-9cbd-efdb1a67b855 tempest-MigrationsAdminTest-1582320444 tempest-MigrationsAdminTest-1582320444-project-member] Lock "cf4e241c-a4e3-438c-8059-c7f7cb41cf45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.564s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.973379] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d671af91-21f0-42a7-8000-c952d6029079 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.980973] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274f5ca4-f3d7-4cfa-a0df-d090e5b88b15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.012061] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae29a51-a08f-4209-bc17-295d4ca23254 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.019670] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4456d5e1-e7ac-47de-ac65-cc240b992fd9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.033179] env[61006]: DEBUG nova.compute.provider_tree [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.172680] env[61006]: INFO nova.compute.manager [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 5e4a8c02-be44-4139-b591-e1ed4ef52261] Took 1.03 seconds to deallocate network for instance. [ 689.182214] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.539191] env[61006]: DEBUG nova.scheduler.client.report [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.707427] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.046312] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.047167] env[61006]: ERROR nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] Traceback (most recent call last): [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.driver.spawn(context, instance, image_meta, [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] vm_ref = self.build_virtual_machine(instance, [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.047167] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] for vif in network_info: [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return self._sync_wrapper(fn, *args, **kwargs) [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.wait() [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self[:] = self._gt.wait() [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return self._exit_event.wait() [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] current.throw(*self._exc) [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.047707] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] result = function(*args, **kwargs) [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] return func(*args, **kwargs) [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise e [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] nwinfo = self.network_api.allocate_for_instance( [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] created_port_ids = self._update_ports_for_instance( [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] with excutils.save_and_reraise_exception(): [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] self.force_reraise() [ 690.048143] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise self.value [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] updated_port = self._update_port( [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] _ensure_no_port_binding_failure(port) [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] raise exception.PortBindingFailed(port_id=port['id']) [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] nova.exception.PortBindingFailed: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. [ 690.048498] env[61006]: ERROR nova.compute.manager [instance: 09646374-a3e3-4fc6-a227-950966144315] [ 690.048498] env[61006]: DEBUG nova.compute.utils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.049191] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.244s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.049424] env[61006]: DEBUG nova.objects.instance [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lazy-loading 'resources' on Instance uuid 47124964-8057-46ba-8c57-1a9f4471402c {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 690.050684] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Build of instance 09646374-a3e3-4fc6-a227-950966144315 was re-scheduled: Binding failed for port 78908742-e120-4566-9bee-06b6cc15a8f3, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.051118] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.051396] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquiring lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.051535] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Acquired lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.051684] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.203722] env[61006]: INFO nova.scheduler.client.report [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Deleted allocations for instance 5e4a8c02-be44-4139-b591-e1ed4ef52261 [ 690.591427] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.700281] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.719674] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ab7292ac-ac8c-4131-a258-1189da7f2f26 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "5e4a8c02-be44-4139-b591-e1ed4ef52261" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.631s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.038625] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2609ab-d3cb-43e8-ad4b-db140394edb8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.046710] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1baae72-1fcd-4929-ab33-8d36f3abcfce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.078254] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef50ed5-abb8-4711-bf0e-63cbc6ef6227 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.086673] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded6b6a3-10d8-4ef1-8326-bff9f5ddde7f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.103675] env[61006]: DEBUG nova.compute.provider_tree [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.205569] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Releasing lock "refresh_cache-09646374-a3e3-4fc6-a227-950966144315" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.205863] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.206023] env[61006]: DEBUG nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.206197] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.222930] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.233368] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.607455] env[61006]: DEBUG nova.scheduler.client.report [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.735713] env[61006]: DEBUG nova.network.neutron [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.750508] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.114929] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.065s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.117185] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.052s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.147887] env[61006]: INFO nova.scheduler.client.report [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleted allocations for instance 47124964-8057-46ba-8c57-1a9f4471402c [ 692.238564] env[61006]: INFO nova.compute.manager [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] [instance: 09646374-a3e3-4fc6-a227-950966144315] Took 1.03 seconds to deallocate network for instance. [ 692.514240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "2a4089c9-4229-40bd-8d0e-706bba94655f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.514240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.657927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c801caa9-7d4a-419b-9f22-8a61a3d28422 tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "47124964-8057-46ba-8c57-1a9f4471402c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.569s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.025265] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df94a1a0-def1-4c87-abf9-fc3ad3dacfa1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.032830] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26cc5a2-7c0e-42df-bcf4-ef34ef1df7f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.064178] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff2e11e-9406-41da-90dd-1bd16ac2f684 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.071556] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785019b0-dba9-4d97-8061-bda706a95f75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.085213] env[61006]: DEBUG nova.compute.provider_tree [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.289490] env[61006]: INFO nova.scheduler.client.report [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Deleted allocations for instance 09646374-a3e3-4fc6-a227-950966144315 [ 693.588854] env[61006]: DEBUG nova.scheduler.client.report [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.780066] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.780441] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.780651] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.780832] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.780991] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.783344] env[61006]: INFO nova.compute.manager [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Terminating instance [ 693.785430] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "refresh_cache-6642ffcb-cb01-4e38-a27c-bf4e4c938a17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.785584] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquired lock "refresh_cache-6642ffcb-cb01-4e38-a27c-bf4e4c938a17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.785743] env[61006]: DEBUG nova.network.neutron [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.801485] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f5d1c1d4-4da3-4859-a54c-096073a003f5 tempest-ListImageFiltersTestJSON-786256351 tempest-ListImageFiltersTestJSON-786256351-project-member] Lock "09646374-a3e3-4fc6-a227-950966144315" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.757s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.094982] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.095709] env[61006]: ERROR nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Traceback (most recent call last): [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.driver.spawn(context, instance, image_meta, [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] vm_ref = self.build_virtual_machine(instance, [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.095709] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] for vif in network_info: [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self._sync_wrapper(fn, *args, **kwargs) [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.wait() [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self[:] = self._gt.wait() [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self._exit_event.wait() [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] result = hub.switch() [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.095996] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return self.greenlet.switch() [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] result = function(*args, **kwargs) [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] return func(*args, **kwargs) [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise e [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] nwinfo = self.network_api.allocate_for_instance( [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] created_port_ids = self._update_ports_for_instance( [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] with excutils.save_and_reraise_exception(): [ 694.096322] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] self.force_reraise() [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise self.value [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] updated_port = self._update_port( [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] _ensure_no_port_binding_failure(port) [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] raise exception.PortBindingFailed(port_id=port['id']) [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] nova.exception.PortBindingFailed: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. [ 694.096643] env[61006]: ERROR nova.compute.manager [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] [ 694.096914] env[61006]: DEBUG nova.compute.utils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.098564] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Build of instance 8102112b-ee4d-4c7e-9b67-9b6a9c032f56 was re-scheduled: Binding failed for port 430bf9f1-f7f5-431f-ad3a-9fb22b8dbc01, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.099095] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.099397] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.099642] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.099904] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.101126] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.528s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.102716] env[61006]: INFO nova.compute.claims [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.304017] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.313518] env[61006]: DEBUG nova.network.neutron [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.401889] env[61006]: DEBUG nova.network.neutron [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.638196] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.791780] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.835267] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.904079] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Releasing lock "refresh_cache-6642ffcb-cb01-4e38-a27c-bf4e4c938a17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.904521] env[61006]: DEBUG nova.compute.manager [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.904716] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.905655] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111eac80-85dc-4233-bcaa-f3fd9bd3e973 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.917384] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 694.917384] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13c4968d-53ae-4e7d-aaed-1652a00d23af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.921087] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 694.921087] env[61006]: value = "task-1336909" [ 694.921087] env[61006]: _type = "Task" [ 694.921087] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.930659] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336909, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.294121] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Releasing lock "refresh_cache-8102112b-ee4d-4c7e-9b67-9b6a9c032f56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.294361] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 695.294544] env[61006]: DEBUG nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.294708] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.320713] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.435810] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336909, 'name': PowerOffVM_Task, 'duration_secs': 0.118783} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.436117] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 695.436590] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 695.436883] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11f598bb-94b7-4c45-8b59-675bd2216129 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.468680] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 695.468805] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 695.468976] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleting the datastore file [datastore2] 6642ffcb-cb01-4e38-a27c-bf4e4c938a17 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.469257] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b01222a6-152e-40ff-84ce-4347104c5f87 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.478009] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for the task: (returnval){ [ 695.478009] env[61006]: value = "task-1336911" [ 695.478009] env[61006]: _type = "Task" [ 695.478009] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.488406] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336911, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.598482] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d280c8-1f62-48e4-899d-209155506983 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.606096] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2836e89a-d539-4431-8479-98b34bbd1832 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.638433] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cdebfb-1e3b-46ba-bbe3-90f4d83babb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.645892] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137c9d2f-4718-4db5-a8c3-8d72dab32a5b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.659544] env[61006]: DEBUG nova.compute.provider_tree [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.828244] env[61006]: DEBUG nova.network.neutron [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.990860] env[61006]: DEBUG oslo_vmware.api [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Task: {'id': task-1336911, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.250363} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.991155] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 695.991373] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 695.991949] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.991949] env[61006]: INFO nova.compute.manager [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Took 1.09 seconds to destroy the instance on the hypervisor. [ 695.991949] env[61006]: DEBUG oslo.service.loopingcall [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.992155] env[61006]: DEBUG nova.compute.manager [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.992251] env[61006]: DEBUG nova.network.neutron [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.012440] env[61006]: DEBUG nova.network.neutron [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.165122] env[61006]: DEBUG nova.scheduler.client.report [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.330059] env[61006]: INFO nova.compute.manager [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 8102112b-ee4d-4c7e-9b67-9b6a9c032f56] Took 1.04 seconds to deallocate network for instance. [ 696.517267] env[61006]: DEBUG nova.network.neutron [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.673053] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.673053] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 696.678044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.102s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.678893] env[61006]: INFO nova.compute.claims [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.022054] env[61006]: INFO nova.compute.manager [-] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Took 1.03 seconds to deallocate network for instance. [ 697.192247] env[61006]: DEBUG nova.compute.utils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.197021] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.197021] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.253983] env[61006]: DEBUG nova.policy [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd11ae452bbb84fe1896d56ccfbb0a8ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f94fffee6381498d88c9509b2117ee71', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.371305] env[61006]: INFO nova.scheduler.client.report [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Deleted allocations for instance 8102112b-ee4d-4c7e-9b67-9b6a9c032f56 [ 697.540704] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.659466] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Successfully created port: 6fbf4c93-f06e-4138-ab07-95549266ba2e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.703583] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.884447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-44f0c961-23ec-4cef-ba0e-acb06e564d5f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "8102112b-ee4d-4c7e-9b67-9b6a9c032f56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.668s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.125132] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2db8691-8bc7-4613-9f6d-9212d97b222f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.134484] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a23fd7d-e6ca-4098-a7a4-a98b50585e6a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.169636] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f81a2c-a6c2-46ca-9b70-0cddc92ccc7f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.179170] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cefbe6-be7f-4590-a1ae-c4dc067547ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.193398] env[61006]: DEBUG nova.compute.provider_tree [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.217789] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "42b92d52-d1f0-48ff-94b6-6164b387456c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.218365] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.387641] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 698.697316] env[61006]: DEBUG nova.scheduler.client.report [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.716625] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 698.744711] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.744946] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.745107] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.745285] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.745428] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.745565] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.745764] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.745916] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.746086] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.746244] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.746411] env[61006]: DEBUG nova.virt.hardware [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.747288] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836700fa-f46b-4db7-8b43-e409a72d2ac7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.756134] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cd3659-428a-4aa8-9139-cdf7251f2462 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.914629] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.063605] env[61006]: ERROR nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 699.063605] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.063605] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.063605] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.063605] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.063605] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.063605] env[61006]: ERROR nova.compute.manager raise self.value [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.063605] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.063605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.063605] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.065617] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.065617] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.065617] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 699.065617] env[61006]: ERROR nova.compute.manager [ 699.065617] env[61006]: Traceback (most recent call last): [ 699.065617] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.065617] env[61006]: listener.cb(fileno) [ 699.065617] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.065617] env[61006]: result = function(*args, **kwargs) [ 699.065617] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.065617] env[61006]: return func(*args, **kwargs) [ 699.065617] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.065617] env[61006]: raise e [ 699.065617] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.065617] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 699.065617] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.065617] env[61006]: created_port_ids = self._update_ports_for_instance( [ 699.065617] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.065617] env[61006]: with excutils.save_and_reraise_exception(): [ 699.065617] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.065617] env[61006]: self.force_reraise() [ 699.065617] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.065617] env[61006]: raise self.value [ 699.065617] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.065617] env[61006]: updated_port = self._update_port( [ 699.065617] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.065617] env[61006]: _ensure_no_port_binding_failure(port) [ 699.065617] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.065617] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.066416] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 699.066416] env[61006]: Removing descriptor: 20 [ 699.066416] env[61006]: ERROR nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Traceback (most recent call last): [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] yield resources [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.driver.spawn(context, instance, image_meta, [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.066416] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] vm_ref = self.build_virtual_machine(instance, [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] for vif in network_info: [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self._sync_wrapper(fn, *args, **kwargs) [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.wait() [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self[:] = self._gt.wait() [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self._exit_event.wait() [ 699.066731] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] result = hub.switch() [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self.greenlet.switch() [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] result = function(*args, **kwargs) [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return func(*args, **kwargs) [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise e [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] nwinfo = self.network_api.allocate_for_instance( [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.067058] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] created_port_ids = self._update_ports_for_instance( [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] with excutils.save_and_reraise_exception(): [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.force_reraise() [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise self.value [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] updated_port = self._update_port( [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] _ensure_no_port_binding_failure(port) [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.067387] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise exception.PortBindingFailed(port_id=port['id']) [ 699.067675] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 699.067675] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] [ 699.067675] env[61006]: INFO nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Terminating instance [ 699.067675] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.067675] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquired lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.067675] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.203705] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.204687] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.207749] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.676s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.212165] env[61006]: INFO nova.compute.claims [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.255246] env[61006]: DEBUG nova.compute.manager [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Received event network-changed-6fbf4c93-f06e-4138-ab07-95549266ba2e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.255246] env[61006]: DEBUG nova.compute.manager [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Refreshing instance network info cache due to event network-changed-6fbf4c93-f06e-4138-ab07-95549266ba2e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 699.255246] env[61006]: DEBUG oslo_concurrency.lockutils [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] Acquiring lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.587312] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.701181] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.707632] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "598c0d72-d679-49a8-b17c-f5f341c205e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.707859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.717161] env[61006]: DEBUG nova.compute.utils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.723023] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.723023] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.740797] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "29c41817-2189-4622-8f35-86f61eb34bed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.741075] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.779525] env[61006]: DEBUG nova.policy [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd11ae452bbb84fe1896d56ccfbb0a8ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f94fffee6381498d88c9509b2117ee71', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.845933] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.845933] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.049922] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Successfully created port: a5b9458d-b5b2-4340-b44b-a237ece911cc {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.204471] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Releasing lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.205462] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.205644] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.206221] env[61006]: DEBUG oslo_concurrency.lockutils [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] Acquired lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.206410] env[61006]: DEBUG nova.network.neutron [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Refreshing network info cache for port 6fbf4c93-f06e-4138-ab07-95549266ba2e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.207792] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb114b73-a9d6-4686-88dc-4b76544f1eb8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.219652] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec77f5e9-c56f-4b7a-9561-e146cdd86984 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.233090] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.247052] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0b97c94-a055-4ce3-b961-1cdb6a5955a3 could not be found. [ 700.247302] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.247703] env[61006]: INFO nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.247703] env[61006]: DEBUG oslo.service.loopingcall [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.248214] env[61006]: DEBUG nova.compute.manager [-] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.248214] env[61006]: DEBUG nova.network.neutron [-] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.267521] env[61006]: DEBUG nova.network.neutron [-] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.725590] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e539e89-004a-4315-9b45-be2c9354ee0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.735211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1400b1b-f622-4411-93fb-98c2f7eddb11 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.740827] env[61006]: DEBUG nova.network.neutron [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.778324] env[61006]: DEBUG nova.network.neutron [-] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.782150] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcda950-f35b-4389-acaf-a38f0bcc51a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.789987] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3bd808-27e2-409f-9df0-40aa5d468b3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.806785] env[61006]: DEBUG nova.compute.provider_tree [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.917810] env[61006]: DEBUG nova.network.neutron [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.178894] env[61006]: ERROR nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 701.178894] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.178894] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.178894] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.178894] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.178894] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.178894] env[61006]: ERROR nova.compute.manager raise self.value [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.178894] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.178894] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.178894] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.179448] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.179448] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.179448] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 701.179448] env[61006]: ERROR nova.compute.manager [ 701.179448] env[61006]: Traceback (most recent call last): [ 701.179448] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.179448] env[61006]: listener.cb(fileno) [ 701.179448] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.179448] env[61006]: result = function(*args, **kwargs) [ 701.179448] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.179448] env[61006]: return func(*args, **kwargs) [ 701.179448] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.179448] env[61006]: raise e [ 701.179448] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.179448] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 701.179448] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.179448] env[61006]: created_port_ids = self._update_ports_for_instance( [ 701.179448] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.179448] env[61006]: with excutils.save_and_reraise_exception(): [ 701.179448] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.179448] env[61006]: self.force_reraise() [ 701.179448] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.179448] env[61006]: raise self.value [ 701.179448] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.179448] env[61006]: updated_port = self._update_port( [ 701.179448] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.179448] env[61006]: _ensure_no_port_binding_failure(port) [ 701.179448] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.179448] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.180324] env[61006]: nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 701.180324] env[61006]: Removing descriptor: 20 [ 701.251146] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.279448] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.279645] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.279772] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.279948] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.280100] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.280260] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.280492] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.280648] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.280807] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.280969] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.281191] env[61006]: DEBUG nova.virt.hardware [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.282013] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4bcf24-0953-474b-ae9f-f7ffe4eb0f5f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.286222] env[61006]: INFO nova.compute.manager [-] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Took 1.04 seconds to deallocate network for instance. [ 701.287510] env[61006]: DEBUG nova.compute.manager [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Received event network-vif-deleted-6fbf4c93-f06e-4138-ab07-95549266ba2e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.287831] env[61006]: DEBUG nova.compute.manager [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Received event network-changed-a5b9458d-b5b2-4340-b44b-a237ece911cc {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.287831] env[61006]: DEBUG nova.compute.manager [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Refreshing instance network info cache due to event network-changed-a5b9458d-b5b2-4340-b44b-a237ece911cc. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.288065] env[61006]: DEBUG oslo_concurrency.lockutils [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] Acquiring lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.288237] env[61006]: DEBUG oslo_concurrency.lockutils [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] Acquired lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.288497] env[61006]: DEBUG nova.network.neutron [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Refreshing network info cache for port a5b9458d-b5b2-4340-b44b-a237ece911cc {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.291355] env[61006]: DEBUG nova.compute.claims [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.291510] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.296270] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0438d5be-9dd0-4bd4-804d-9997b56c5cc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.310994] env[61006]: DEBUG nova.scheduler.client.report [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.314903] env[61006]: ERROR nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] Traceback (most recent call last): [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] yield resources [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.driver.spawn(context, instance, image_meta, [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] vm_ref = self.build_virtual_machine(instance, [ 701.314903] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] for vif in network_info: [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return self._sync_wrapper(fn, *args, **kwargs) [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.wait() [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self[:] = self._gt.wait() [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return self._exit_event.wait() [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.315508] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] current.throw(*self._exc) [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] result = function(*args, **kwargs) [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return func(*args, **kwargs) [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise e [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] nwinfo = self.network_api.allocate_for_instance( [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] created_port_ids = self._update_ports_for_instance( [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] with excutils.save_and_reraise_exception(): [ 701.315872] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.force_reraise() [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise self.value [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] updated_port = self._update_port( [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] _ensure_no_port_binding_failure(port) [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise exception.PortBindingFailed(port_id=port['id']) [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 701.316205] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] [ 701.316205] env[61006]: INFO nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Terminating instance [ 701.317710] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.420784] env[61006]: DEBUG oslo_concurrency.lockutils [req-2e56ad25-6966-40c9-928d-1a750edb9119 req-65116042-7390-4b3a-9681-5084f54d70fc service nova] Releasing lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.816184] env[61006]: DEBUG nova.network.neutron [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.819238] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.819510] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.822256] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.869s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.934824] env[61006]: DEBUG nova.network.neutron [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.331556] env[61006]: DEBUG nova.compute.utils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.333765] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.333765] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 702.372441] env[61006]: DEBUG nova.policy [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9f4d8f48dc44e3490ac29feab6c11d7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '82fea87577594087a0ca1f08fd8f95af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.439853] env[61006]: DEBUG oslo_concurrency.lockutils [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] Releasing lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.440132] env[61006]: DEBUG nova.compute.manager [req-a41ac513-df37-4169-a7d4-a97e7f9cedda req-e0f71868-b8b3-4da1-9f4e-f84aa31d5e3b service nova] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Received event network-vif-deleted-a5b9458d-b5b2-4340-b44b-a237ece911cc {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.440934] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquired lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.440934] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.741528] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Successfully created port: fbadc913-f41e-4f04-b065-9052f88f9c4d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.750794] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae590477-2a1b-4318-b278-8db7bc936857 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.758825] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01083969-01aa-4484-8e26-4f8869bb0fb0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.790876] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb027ddc-c27f-49a9-943b-f7ec53d275cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.798370] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c4edef-459f-4679-93d4-b8fadb5d349d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.811468] env[61006]: DEBUG nova.compute.provider_tree [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.836162] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.964293] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.104469] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.314249] env[61006]: DEBUG nova.scheduler.client.report [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.578898] env[61006]: DEBUG nova.compute.manager [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Received event network-changed-fbadc913-f41e-4f04-b065-9052f88f9c4d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.579176] env[61006]: DEBUG nova.compute.manager [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Refreshing instance network info cache due to event network-changed-fbadc913-f41e-4f04-b065-9052f88f9c4d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.579472] env[61006]: DEBUG oslo_concurrency.lockutils [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] Acquiring lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.579668] env[61006]: DEBUG oslo_concurrency.lockutils [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] Acquired lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.579877] env[61006]: DEBUG nova.network.neutron [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Refreshing network info cache for port fbadc913-f41e-4f04-b065-9052f88f9c4d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.608172] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Releasing lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.608591] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.608778] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.609058] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f276b52b-ddef-4392-bfa7-8bc1c2d1126b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.618592] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804fcb38-e3bb-4b91-b1c3-4e6e02bcaaf6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.643753] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7325d28-ed07-4dab-b440-2daf766882de could not be found. [ 703.643970] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.644160] env[61006]: INFO nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.644398] env[61006]: DEBUG oslo.service.loopingcall [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.644611] env[61006]: DEBUG nova.compute.manager [-] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.644704] env[61006]: DEBUG nova.network.neutron [-] [instance: f7325d28-ed07-4dab-b440-2daf766882de] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.660762] env[61006]: DEBUG nova.network.neutron [-] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.757725] env[61006]: ERROR nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 703.757725] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.757725] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.757725] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.757725] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.757725] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.757725] env[61006]: ERROR nova.compute.manager raise self.value [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.757725] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.757725] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.757725] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.758226] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.758226] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.758226] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 703.758226] env[61006]: ERROR nova.compute.manager [ 703.758226] env[61006]: Traceback (most recent call last): [ 703.758226] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.758226] env[61006]: listener.cb(fileno) [ 703.758226] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.758226] env[61006]: result = function(*args, **kwargs) [ 703.758226] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.758226] env[61006]: return func(*args, **kwargs) [ 703.758226] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.758226] env[61006]: raise e [ 703.758226] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.758226] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 703.758226] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.758226] env[61006]: created_port_ids = self._update_ports_for_instance( [ 703.758226] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.758226] env[61006]: with excutils.save_and_reraise_exception(): [ 703.758226] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.758226] env[61006]: self.force_reraise() [ 703.758226] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.758226] env[61006]: raise self.value [ 703.758226] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.758226] env[61006]: updated_port = self._update_port( [ 703.758226] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.758226] env[61006]: _ensure_no_port_binding_failure(port) [ 703.758226] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.758226] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.758994] env[61006]: nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 703.758994] env[61006]: Removing descriptor: 20 [ 703.819622] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.820400] env[61006]: ERROR nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Traceback (most recent call last): [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.driver.spawn(context, instance, image_meta, [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] vm_ref = self.build_virtual_machine(instance, [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.820400] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] for vif in network_info: [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return self._sync_wrapper(fn, *args, **kwargs) [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.wait() [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self[:] = self._gt.wait() [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return self._exit_event.wait() [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] current.throw(*self._exc) [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.820792] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] result = function(*args, **kwargs) [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] return func(*args, **kwargs) [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise e [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] nwinfo = self.network_api.allocate_for_instance( [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] created_port_ids = self._update_ports_for_instance( [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] with excutils.save_and_reraise_exception(): [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] self.force_reraise() [ 703.821185] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise self.value [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] updated_port = self._update_port( [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] _ensure_no_port_binding_failure(port) [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] raise exception.PortBindingFailed(port_id=port['id']) [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] nova.exception.PortBindingFailed: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. [ 703.821550] env[61006]: ERROR nova.compute.manager [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] [ 703.821550] env[61006]: DEBUG nova.compute.utils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.822216] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.592s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.825656] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Build of instance 5a2ddd3d-c939-4f20-a40d-108e805d64c4 was re-scheduled: Binding failed for port 2d065202-2687-4ff8-87c4-adbb2b7099d8, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.826927] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.826927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquiring lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.826927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Acquired lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.826927] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.845767] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.873570] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.873807] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.873956] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.874156] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.874295] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.874437] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.874629] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.874779] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.874937] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.876039] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.876039] env[61006]: DEBUG nova.virt.hardware [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.876414] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23b49f6-f438-476d-9874-e9eb25fb59f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.885280] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae546c21-7595-470f-be04-9685910a2659 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.900279] env[61006]: ERROR nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Traceback (most recent call last): [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] yield resources [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.driver.spawn(context, instance, image_meta, [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] vm_ref = self.build_virtual_machine(instance, [ 703.900279] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] for vif in network_info: [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return self._sync_wrapper(fn, *args, **kwargs) [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.wait() [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self[:] = self._gt.wait() [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return self._exit_event.wait() [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.900633] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] current.throw(*self._exc) [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] result = function(*args, **kwargs) [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return func(*args, **kwargs) [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise e [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] nwinfo = self.network_api.allocate_for_instance( [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] created_port_ids = self._update_ports_for_instance( [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] with excutils.save_and_reraise_exception(): [ 703.901072] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.force_reraise() [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise self.value [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] updated_port = self._update_port( [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] _ensure_no_port_binding_failure(port) [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise exception.PortBindingFailed(port_id=port['id']) [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 703.901449] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] [ 703.901449] env[61006]: INFO nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Terminating instance [ 703.903455] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquiring lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.098764] env[61006]: DEBUG nova.network.neutron [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.163588] env[61006]: DEBUG nova.network.neutron [-] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.187722] env[61006]: DEBUG nova.network.neutron [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.345859] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.419292] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.667066] env[61006]: INFO nova.compute.manager [-] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Took 1.02 seconds to deallocate network for instance. [ 704.669190] env[61006]: DEBUG nova.compute.claims [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.669374] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.690382] env[61006]: DEBUG oslo_concurrency.lockutils [req-f2e3a905-4c1e-483a-8c70-69c4524f4a42 req-49c86d9e-44eb-421d-b690-6dd85aedcadc service nova] Releasing lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.690928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquired lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.691119] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.697473] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993610e0-45db-4ea5-9cc5-e6f442bcabee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.705625] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee641da0-6f56-4563-86af-28b866474418 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.736412] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d77860-dbaf-45c6-9e56-c0cea148b15a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.743715] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbb9a3a-629d-4817-9341-1d7c0fc46975 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.756999] env[61006]: DEBUG nova.compute.provider_tree [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.922172] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Releasing lock "refresh_cache-5a2ddd3d-c939-4f20-a40d-108e805d64c4" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.922439] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.922598] env[61006]: DEBUG nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.922760] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.945768] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.219060] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.259949] env[61006]: DEBUG nova.scheduler.client.report [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.305759] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.448433] env[61006]: DEBUG nova.network.neutron [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.610586] env[61006]: DEBUG nova.compute.manager [req-90d0c08d-247f-4eca-8efb-bf4b42b26d97 req-b2a219f5-ce6c-461b-818d-a19bf55eeb95 service nova] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Received event network-vif-deleted-fbadc913-f41e-4f04-b065-9052f88f9c4d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.765431] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.766130] env[61006]: ERROR nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Traceback (most recent call last): [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.driver.spawn(context, instance, image_meta, [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] vm_ref = self.build_virtual_machine(instance, [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.766130] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] for vif in network_info: [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self._sync_wrapper(fn, *args, **kwargs) [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.wait() [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self[:] = self._gt.wait() [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self._exit_event.wait() [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] result = hub.switch() [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.766439] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return self.greenlet.switch() [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] result = function(*args, **kwargs) [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] return func(*args, **kwargs) [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise e [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] nwinfo = self.network_api.allocate_for_instance( [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] created_port_ids = self._update_ports_for_instance( [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] with excutils.save_and_reraise_exception(): [ 705.766753] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] self.force_reraise() [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise self.value [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] updated_port = self._update_port( [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] _ensure_no_port_binding_failure(port) [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] raise exception.PortBindingFailed(port_id=port['id']) [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] nova.exception.PortBindingFailed: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. [ 705.767135] env[61006]: ERROR nova.compute.manager [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] [ 705.767442] env[61006]: DEBUG nova.compute.utils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.768182] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.984s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.769624] env[61006]: INFO nova.compute.claims [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.772236] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Build of instance 1542f4da-16bb-47c3-b11a-a9ac332ac917 was re-scheduled: Binding failed for port 6f73dae8-091b-49cc-902a-3aa360a5d6ca, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.772704] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.772928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquiring lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.773085] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Acquired lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.773243] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.811030] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Releasing lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.811030] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.811030] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.811030] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fce71d10-b887-4e4d-9839-ffb391951220 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.818430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87e43b2-cece-4c45-943c-d5623e834649 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.839396] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5525a859-eb09-45a6-80d6-8a466976a7bf could not be found. [ 705.839607] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.839782] env[61006]: INFO nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 705.840035] env[61006]: DEBUG oslo.service.loopingcall [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.840265] env[61006]: DEBUG nova.compute.manager [-] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.840380] env[61006]: DEBUG nova.network.neutron [-] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.854041] env[61006]: DEBUG nova.network.neutron [-] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.951244] env[61006]: INFO nova.compute.manager [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] [instance: 5a2ddd3d-c939-4f20-a40d-108e805d64c4] Took 1.03 seconds to deallocate network for instance. [ 706.293707] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.350804] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.356580] env[61006]: DEBUG nova.network.neutron [-] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.854981] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Releasing lock "refresh_cache-1542f4da-16bb-47c3-b11a-a9ac332ac917" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.854981] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.854981] env[61006]: DEBUG nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.854981] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.858721] env[61006]: INFO nova.compute.manager [-] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Took 1.02 seconds to deallocate network for instance. [ 706.861839] env[61006]: DEBUG nova.compute.claims [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.862308] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.869089] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.981020] env[61006]: INFO nova.scheduler.client.report [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Deleted allocations for instance 5a2ddd3d-c939-4f20-a40d-108e805d64c4 [ 707.179065] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40567dd-ab74-4893-a76f-cbdebef72fdb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.185932] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531671e4-050e-483c-b13a-3fd188ba9cc0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.224394] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1b673e-6bcf-4138-99d7-8a8783be0374 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.232562] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978c9b02-8958-4de1-bd48-25d48f39c6b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.249614] env[61006]: DEBUG nova.compute.provider_tree [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.372323] env[61006]: DEBUG nova.network.neutron [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.491351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1e7a1163-3a76-486a-a102-318f504c021d tempest-ServersTestBootFromVolume-62526034 tempest-ServersTestBootFromVolume-62526034-project-member] Lock "5a2ddd3d-c939-4f20-a40d-108e805d64c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.476s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.753800] env[61006]: DEBUG nova.scheduler.client.report [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.874921] env[61006]: INFO nova.compute.manager [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] [instance: 1542f4da-16bb-47c3-b11a-a9ac332ac917] Took 1.02 seconds to deallocate network for instance. [ 707.994091] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.259086] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.261041] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.262241] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.556s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.264093] env[61006]: INFO nova.compute.claims [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.518620] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.768307] env[61006]: DEBUG nova.compute.utils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.771619] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.771868] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 708.852013] env[61006]: DEBUG nova.policy [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f0776aaae4b47429ed7b02edb82858d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92dfee95ecf34a9e8174f4395eda4973', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.916537] env[61006]: INFO nova.scheduler.client.report [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Deleted allocations for instance 1542f4da-16bb-47c3-b11a-a9ac332ac917 [ 709.273458] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.346165] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Successfully created port: 31e7e06b-bbd4-4886-b54a-e373e0347595 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.424457] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99b827e6-4510-4b59-94c8-66d0a70f53f4 tempest-TenantUsagesTestJSON-1966759453 tempest-TenantUsagesTestJSON-1966759453-project-member] Lock "1542f4da-16bb-47c3-b11a-a9ac332ac917" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.270s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.800062] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4969f220-d14e-4655-a9bb-abf4232b842c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.807337] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161107b2-0721-4226-9b08-d4e9e73f0e7b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.842059] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d2e37e-6db4-4ac9-829c-245029940b92 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.851871] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64f7c53-7018-488f-a0fb-83205e634eaf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.867010] env[61006]: DEBUG nova.compute.provider_tree [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.926896] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.295203] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.334202] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.334457] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.334610] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.334790] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.334970] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.343056] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.343056] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.343056] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.343056] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.343056] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.343554] env[61006]: DEBUG nova.virt.hardware [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.343554] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1995113f-de25-4372-ab79-111255666279 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.353038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c910eed7-1ae3-4f1f-a392-c1dabc748a63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.369189] env[61006]: DEBUG nova.compute.manager [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Received event network-changed-31e7e06b-bbd4-4886-b54a-e373e0347595 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.369189] env[61006]: DEBUG nova.compute.manager [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Refreshing instance network info cache due to event network-changed-31e7e06b-bbd4-4886-b54a-e373e0347595. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.369189] env[61006]: DEBUG oslo_concurrency.lockutils [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] Acquiring lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.369189] env[61006]: DEBUG oslo_concurrency.lockutils [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] Acquired lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.369189] env[61006]: DEBUG nova.network.neutron [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Refreshing network info cache for port 31e7e06b-bbd4-4886-b54a-e373e0347595 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.371904] env[61006]: DEBUG nova.scheduler.client.report [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.458684] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.518808] env[61006]: ERROR nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 710.518808] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.518808] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.518808] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.518808] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.518808] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.518808] env[61006]: ERROR nova.compute.manager raise self.value [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.518808] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.518808] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.518808] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.519328] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.519328] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.519328] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 710.519328] env[61006]: ERROR nova.compute.manager [ 710.519328] env[61006]: Traceback (most recent call last): [ 710.519328] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.519328] env[61006]: listener.cb(fileno) [ 710.519328] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.519328] env[61006]: result = function(*args, **kwargs) [ 710.519328] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.519328] env[61006]: return func(*args, **kwargs) [ 710.519328] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.519328] env[61006]: raise e [ 710.519328] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.519328] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 710.519328] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.519328] env[61006]: created_port_ids = self._update_ports_for_instance( [ 710.519328] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.519328] env[61006]: with excutils.save_and_reraise_exception(): [ 710.519328] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.519328] env[61006]: self.force_reraise() [ 710.519328] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.519328] env[61006]: raise self.value [ 710.519328] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.519328] env[61006]: updated_port = self._update_port( [ 710.519328] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.519328] env[61006]: _ensure_no_port_binding_failure(port) [ 710.519328] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.519328] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.520215] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 710.520215] env[61006]: Removing descriptor: 20 [ 710.520215] env[61006]: ERROR nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Traceback (most recent call last): [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] yield resources [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.driver.spawn(context, instance, image_meta, [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.520215] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] vm_ref = self.build_virtual_machine(instance, [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] for vif in network_info: [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self._sync_wrapper(fn, *args, **kwargs) [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.wait() [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self[:] = self._gt.wait() [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self._exit_event.wait() [ 710.520618] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] result = hub.switch() [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self.greenlet.switch() [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] result = function(*args, **kwargs) [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return func(*args, **kwargs) [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise e [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] nwinfo = self.network_api.allocate_for_instance( [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.521055] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] created_port_ids = self._update_ports_for_instance( [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] with excutils.save_and_reraise_exception(): [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.force_reraise() [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise self.value [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] updated_port = self._update_port( [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] _ensure_no_port_binding_failure(port) [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.521711] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise exception.PortBindingFailed(port_id=port['id']) [ 710.522228] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 710.522228] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] [ 710.522228] env[61006]: INFO nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Terminating instance [ 710.522228] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquiring lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.878421] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.616s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.881383] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.881693] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.131s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.883094] env[61006]: INFO nova.compute.claims [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.908733] env[61006]: DEBUG nova.network.neutron [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.981369] env[61006]: DEBUG nova.network.neutron [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.390094] env[61006]: DEBUG nova.compute.utils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.395586] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.396988] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 711.483880] env[61006]: DEBUG oslo_concurrency.lockutils [req-c9a7edbb-a5f0-4703-beeb-fe37bcba3b9b req-a6b3f722-81ad-4f4c-a529-4a2ff6ea6498 service nova] Releasing lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.484342] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquired lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.484502] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.504356] env[61006]: DEBUG nova.policy [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b21cc9de4af44a0fa832dbf9ca1eea1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4affaba7247546af9eecf7dcfd47b25f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.912207] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.017750] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.260883] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.298800] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Successfully created port: 11aee232-a61b-434b-9bed-13c67e38a458 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.381351] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427e6390-10db-4cf5-87ac-4eb569f3f4a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.389961] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f520aaf-a80c-47d7-90eb-951d001a895f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.433236] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaffcd9a-0d75-4ab3-82ba-6c5e58565a4c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.436989] env[61006]: DEBUG nova.compute.manager [req-b72ff254-2388-4081-b149-e2bdaf37cf91 req-2d32a6bf-3e49-498c-baa5-4e69067fed50 service nova] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Received event network-vif-deleted-31e7e06b-bbd4-4886-b54a-e373e0347595 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.444164] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583894bd-f9f5-4913-90cd-2f58cf8a84ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.465030] env[61006]: DEBUG nova.compute.provider_tree [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.764483] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Releasing lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.764659] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.764852] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.765170] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2983f14-a146-490e-ae13-6d7540d285b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.781241] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae66c6e3-e081-4b78-9413-52b20bf96f6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.803298] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6892ecb-0fb8-44da-9362-8806c429626b could not be found. [ 712.803561] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.803748] env[61006]: INFO nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 712.807019] env[61006]: DEBUG oslo.service.loopingcall [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.807019] env[61006]: DEBUG nova.compute.manager [-] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.807019] env[61006]: DEBUG nova.network.neutron [-] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.845064] env[61006]: DEBUG nova.network.neutron [-] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.932925] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.966429] env[61006]: DEBUG nova.scheduler.client.report [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.972048] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.972351] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.972427] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.972586] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.972727] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.972868] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.973723] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.973947] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.974158] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.974328] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.974503] env[61006]: DEBUG nova.virt.hardware [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.975607] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017b209e-dbd7-4aec-bf89-b7551d06ebb0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.985190] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb4a1bd-aa2b-4088-bd1b-93547fdc3544 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.346725] env[61006]: DEBUG nova.network.neutron [-] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.481217] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.481217] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.482964] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.648s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.484493] env[61006]: INFO nova.compute.claims [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.849760] env[61006]: INFO nova.compute.manager [-] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Took 1.04 seconds to deallocate network for instance. [ 713.854056] env[61006]: DEBUG nova.compute.claims [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.854258] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.984017] env[61006]: ERROR nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 713.984017] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.984017] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.984017] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.984017] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.984017] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.984017] env[61006]: ERROR nova.compute.manager raise self.value [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.984017] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 713.984017] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.984017] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 713.984625] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.984625] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 713.984625] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 713.984625] env[61006]: ERROR nova.compute.manager [ 713.984625] env[61006]: Traceback (most recent call last): [ 713.984625] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 713.984625] env[61006]: listener.cb(fileno) [ 713.984625] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.984625] env[61006]: result = function(*args, **kwargs) [ 713.984625] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.984625] env[61006]: return func(*args, **kwargs) [ 713.984625] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.984625] env[61006]: raise e [ 713.984625] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.984625] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 713.984625] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.984625] env[61006]: created_port_ids = self._update_ports_for_instance( [ 713.984625] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.984625] env[61006]: with excutils.save_and_reraise_exception(): [ 713.984625] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.984625] env[61006]: self.force_reraise() [ 713.984625] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.984625] env[61006]: raise self.value [ 713.984625] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.984625] env[61006]: updated_port = self._update_port( [ 713.984625] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.984625] env[61006]: _ensure_no_port_binding_failure(port) [ 713.984625] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.984625] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 713.985804] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 713.985804] env[61006]: Removing descriptor: 20 [ 713.985804] env[61006]: ERROR nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Traceback (most recent call last): [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] yield resources [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.driver.spawn(context, instance, image_meta, [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.985804] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] vm_ref = self.build_virtual_machine(instance, [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] for vif in network_info: [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self._sync_wrapper(fn, *args, **kwargs) [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.wait() [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self[:] = self._gt.wait() [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self._exit_event.wait() [ 713.986348] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] result = hub.switch() [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self.greenlet.switch() [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] result = function(*args, **kwargs) [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return func(*args, **kwargs) [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise e [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] nwinfo = self.network_api.allocate_for_instance( [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.986904] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] created_port_ids = self._update_ports_for_instance( [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] with excutils.save_and_reraise_exception(): [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.force_reraise() [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise self.value [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] updated_port = self._update_port( [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] _ensure_no_port_binding_failure(port) [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.987646] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise exception.PortBindingFailed(port_id=port['id']) [ 713.988535] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 713.988535] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] [ 713.988535] env[61006]: INFO nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Terminating instance [ 713.988773] env[61006]: DEBUG nova.compute.utils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.990676] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquiring lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.990845] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquired lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.991014] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.998919] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.999184] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 714.060515] env[61006]: DEBUG nova.policy [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d13e18f459843429bbc204da327480f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4693570ceab84d1babfa6f8f24447fcd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.472165] env[61006]: DEBUG nova.compute.manager [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Received event network-changed-11aee232-a61b-434b-9bed-13c67e38a458 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.472371] env[61006]: DEBUG nova.compute.manager [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Refreshing instance network info cache due to event network-changed-11aee232-a61b-434b-9bed-13c67e38a458. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 714.472554] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] Acquiring lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.503646] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.531797] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.627122] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Successfully created port: 3e55991e-7041-4159-b3b6-b2acb7def1f2 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.781191] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.097193] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1b571f-2a99-4125-95fd-de9e663e341e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.105909] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c71a22d-85fe-4b76-92f6-ae902347cb2b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.147144] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b540431-3257-4a74-8977-182491c2fee8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.158163] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff55b073-09cb-4869-b9dd-f944c166c5a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.169426] env[61006]: DEBUG nova.compute.provider_tree [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.257799] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Successfully created port: f6c6b67b-a5a7-4476-8362-ebf7b5e0b846 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.292217] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Releasing lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.292217] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 715.292217] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.292217] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] Acquired lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.292217] env[61006]: DEBUG nova.network.neutron [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Refreshing network info cache for port 11aee232-a61b-434b-9bed-13c67e38a458 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 715.292446] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b044970e-1295-4dae-b348-f751ddd04897 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.302609] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cca49cf-8102-4a22-a199-5d62e494edb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.326348] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea2a73fa-3ece-4c88-b117-3b9843a4f224 could not be found. [ 715.326592] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.326790] env[61006]: INFO nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Took 0.04 seconds to destroy the instance on the hypervisor. [ 715.327115] env[61006]: DEBUG oslo.service.loopingcall [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.327370] env[61006]: DEBUG nova.compute.manager [-] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.327510] env[61006]: DEBUG nova.network.neutron [-] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.364262] env[61006]: DEBUG nova.network.neutron [-] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.517233] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.562193] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.563151] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.563151] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.563151] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.563151] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.563310] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.563548] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.563695] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.563875] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.564085] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.564375] env[61006]: DEBUG nova.virt.hardware [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.565130] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca57b857-1bd0-4612-be86-552403eb9c6f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.579193] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d55f6a-b40a-4b6b-b6ae-8a3d1fc5e51d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.673186] env[61006]: DEBUG nova.scheduler.client.report [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.832088] env[61006]: DEBUG nova.network.neutron [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.869442] env[61006]: DEBUG nova.network.neutron [-] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.925081] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Successfully created port: 86e7d7aa-79dd-4d7f-8fcd-e2b8e58334fd {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.028773] env[61006]: DEBUG nova.network.neutron [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.182238] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.696s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.182238] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.183570] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.644s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.184085] env[61006]: DEBUG nova.objects.instance [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lazy-loading 'resources' on Instance uuid 6642ffcb-cb01-4e38-a27c-bf4e4c938a17 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 716.372370] env[61006]: INFO nova.compute.manager [-] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Took 1.04 seconds to deallocate network for instance. [ 716.375143] env[61006]: DEBUG nova.compute.claims [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 716.375264] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.533075] env[61006]: DEBUG oslo_concurrency.lockutils [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] Releasing lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.533393] env[61006]: DEBUG nova.compute.manager [req-c5b08db9-b0f3-42d4-aa0a-c7d73192e8ca req-76403fc0-45a4-488d-a9d2-b49775799d36 service nova] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Received event network-vif-deleted-11aee232-a61b-434b-9bed-13c67e38a458 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.688334] env[61006]: DEBUG nova.compute.utils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.689708] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.689878] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.780333] env[61006]: DEBUG nova.policy [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d7a6da112424c19ab454d35bb576b05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21d42eeeb28455881c03018acccaf62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.895351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.896093] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.020602] env[61006]: DEBUG nova.compute.manager [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Received event network-changed-3e55991e-7041-4159-b3b6-b2acb7def1f2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.020602] env[61006]: DEBUG nova.compute.manager [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Refreshing instance network info cache due to event network-changed-3e55991e-7041-4159-b3b6-b2acb7def1f2. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.020818] env[61006]: DEBUG oslo_concurrency.lockutils [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] Acquiring lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.020964] env[61006]: DEBUG oslo_concurrency.lockutils [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] Acquired lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.021139] env[61006]: DEBUG nova.network.neutron [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Refreshing network info cache for port 3e55991e-7041-4159-b3b6-b2acb7def1f2 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.176479] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf04dae-ad13-460c-b906-5c96142569a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.185320] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e0eaaa-d760-4b4d-84d7-c1189a4abce5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.213243] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.216529] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5374fcb6-ec07-4085-9003-1f90071473d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.225148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7331685d-ef33-4658-b145-5b8e6ca7dad4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.245504] env[61006]: DEBUG nova.compute.provider_tree [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.285231] env[61006]: ERROR nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 717.285231] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.285231] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.285231] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.285231] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.285231] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.285231] env[61006]: ERROR nova.compute.manager raise self.value [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.285231] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.285231] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.285231] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.285759] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.285759] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.285759] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 717.285759] env[61006]: ERROR nova.compute.manager [ 717.285759] env[61006]: Traceback (most recent call last): [ 717.285759] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.285759] env[61006]: listener.cb(fileno) [ 717.285759] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.285759] env[61006]: result = function(*args, **kwargs) [ 717.285759] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.285759] env[61006]: return func(*args, **kwargs) [ 717.285759] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.285759] env[61006]: raise e [ 717.285759] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.285759] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 717.285759] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.285759] env[61006]: created_port_ids = self._update_ports_for_instance( [ 717.285759] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.285759] env[61006]: with excutils.save_and_reraise_exception(): [ 717.285759] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.285759] env[61006]: self.force_reraise() [ 717.285759] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.285759] env[61006]: raise self.value [ 717.285759] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.285759] env[61006]: updated_port = self._update_port( [ 717.285759] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.285759] env[61006]: _ensure_no_port_binding_failure(port) [ 717.285759] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.285759] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.286659] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 717.286659] env[61006]: Removing descriptor: 20 [ 717.286659] env[61006]: ERROR nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Traceback (most recent call last): [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] yield resources [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.driver.spawn(context, instance, image_meta, [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.286659] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] vm_ref = self.build_virtual_machine(instance, [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] for vif in network_info: [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self._sync_wrapper(fn, *args, **kwargs) [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.wait() [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self[:] = self._gt.wait() [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self._exit_event.wait() [ 717.287254] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] result = hub.switch() [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self.greenlet.switch() [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] result = function(*args, **kwargs) [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return func(*args, **kwargs) [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise e [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] nwinfo = self.network_api.allocate_for_instance( [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.287735] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] created_port_ids = self._update_ports_for_instance( [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] with excutils.save_and_reraise_exception(): [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.force_reraise() [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise self.value [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] updated_port = self._update_port( [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] _ensure_no_port_binding_failure(port) [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.288214] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise exception.PortBindingFailed(port_id=port['id']) [ 717.288756] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 717.288756] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] [ 717.288756] env[61006]: INFO nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Terminating instance [ 717.288756] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.292731] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Successfully created port: 374ad675-7064-4009-afd1-7727266d7e86 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.602380] env[61006]: DEBUG nova.network.neutron [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.748346] env[61006]: DEBUG nova.scheduler.client.report [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.770058] env[61006]: DEBUG nova.network.neutron [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.233020] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.256070] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.256535] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.256825] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.257269] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.257662] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.257942] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.259029] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.259029] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.259029] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.259029] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.259029] env[61006]: DEBUG nova.virt.hardware [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.259878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.076s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.262523] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c219000-6475-4665-8087-48d02f9bcec9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.267051] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.353s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.268889] env[61006]: INFO nova.compute.claims [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.275800] env[61006]: DEBUG oslo_concurrency.lockutils [req-71a44be0-4f2a-42f1-bc44-cdedcdad8a8c req-88b52ac1-fc67-40c5-a0e2-26466f846144 service nova] Releasing lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.276242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.276431] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.281812] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfeac83a-4b25-447c-9d8a-d4d3d2100b75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.300428] env[61006]: INFO nova.scheduler.client.report [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Deleted allocations for instance 6642ffcb-cb01-4e38-a27c-bf4e4c938a17 [ 718.478334] env[61006]: ERROR nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 718.478334] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.478334] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.478334] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.478334] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.478334] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.478334] env[61006]: ERROR nova.compute.manager raise self.value [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.478334] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.478334] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.478334] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.479168] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.479168] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.479168] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 718.479168] env[61006]: ERROR nova.compute.manager [ 718.479168] env[61006]: Traceback (most recent call last): [ 718.479168] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.479168] env[61006]: listener.cb(fileno) [ 718.479168] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.479168] env[61006]: result = function(*args, **kwargs) [ 718.479168] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.479168] env[61006]: return func(*args, **kwargs) [ 718.479168] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.479168] env[61006]: raise e [ 718.479168] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.479168] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 718.479168] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.479168] env[61006]: created_port_ids = self._update_ports_for_instance( [ 718.479168] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.479168] env[61006]: with excutils.save_and_reraise_exception(): [ 718.479168] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.479168] env[61006]: self.force_reraise() [ 718.479168] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.479168] env[61006]: raise self.value [ 718.479168] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.479168] env[61006]: updated_port = self._update_port( [ 718.479168] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.479168] env[61006]: _ensure_no_port_binding_failure(port) [ 718.479168] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.479168] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.480962] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 718.480962] env[61006]: Removing descriptor: 17 [ 718.480962] env[61006]: ERROR nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Traceback (most recent call last): [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] yield resources [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.driver.spawn(context, instance, image_meta, [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.480962] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] vm_ref = self.build_virtual_machine(instance, [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] for vif in network_info: [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self._sync_wrapper(fn, *args, **kwargs) [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.wait() [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self[:] = self._gt.wait() [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self._exit_event.wait() [ 718.481593] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] result = hub.switch() [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self.greenlet.switch() [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] result = function(*args, **kwargs) [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return func(*args, **kwargs) [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise e [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] nwinfo = self.network_api.allocate_for_instance( [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.481996] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] created_port_ids = self._update_ports_for_instance( [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] with excutils.save_and_reraise_exception(): [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.force_reraise() [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise self.value [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] updated_port = self._update_port( [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] _ensure_no_port_binding_failure(port) [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.482406] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise exception.PortBindingFailed(port_id=port['id']) [ 718.482856] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 718.482856] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] [ 718.482856] env[61006]: INFO nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Terminating instance [ 718.482856] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquiring lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.482856] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquired lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.482856] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.806884] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5ec58a84-9a95-477a-b9c4-a07dacfc767e tempest-ServerShowV247Test-7899526 tempest-ServerShowV247Test-7899526-project-member] Lock "6642ffcb-cb01-4e38-a27c-bf4e4c938a17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.026s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.819694] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.916110] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.998521] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.058683] env[61006]: DEBUG nova.compute.manager [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Received event network-vif-deleted-3e55991e-7041-4159-b3b6-b2acb7def1f2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.058890] env[61006]: DEBUG nova.compute.manager [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Received event network-changed-374ad675-7064-4009-afd1-7727266d7e86 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.059062] env[61006]: DEBUG nova.compute.manager [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Refreshing instance network info cache due to event network-changed-374ad675-7064-4009-afd1-7727266d7e86. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.059244] env[61006]: DEBUG oslo_concurrency.lockutils [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] Acquiring lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.110493] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.418775] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Releasing lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.419270] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 719.419465] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.422211] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75162ce4-6478-4a78-b659-d04faf9a8cc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.432400] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0143ec01-e3ac-4990-8a7e-32cf01cfe20c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.463522] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34f17493-0d3d-4dfd-a2ac-6adc9b388f95 could not be found. [ 719.463522] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.463522] env[61006]: INFO nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Took 0.04 seconds to destroy the instance on the hypervisor. [ 719.463522] env[61006]: DEBUG oslo.service.loopingcall [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.466709] env[61006]: DEBUG nova.compute.manager [-] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.467490] env[61006]: DEBUG nova.network.neutron [-] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.559951] env[61006]: DEBUG nova.network.neutron [-] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.612858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Releasing lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.613325] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 719.613523] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.613857] env[61006]: DEBUG oslo_concurrency.lockutils [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] Acquired lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.613988] env[61006]: DEBUG nova.network.neutron [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Refreshing network info cache for port 374ad675-7064-4009-afd1-7727266d7e86 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 719.615079] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0bf6bd1-6130-45a8-b9c7-de377c58fa34 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.624711] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c512533-5de8-45af-b11a-dbb3eb295a35 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.651956] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2217133d-52fb-45c4-bb84-63dde0546747 could not be found. [ 719.652316] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.652595] env[61006]: INFO nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Took 0.04 seconds to destroy the instance on the hypervisor. [ 719.652945] env[61006]: DEBUG oslo.service.loopingcall [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.655683] env[61006]: DEBUG nova.compute.manager [-] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.657959] env[61006]: DEBUG nova.network.neutron [-] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.687300] env[61006]: DEBUG nova.network.neutron [-] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.709808] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb78cf9-8f6c-44a4-bd81-de5e55035c5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.717673] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1807086e-95d9-4de7-8a7e-d9972fbb440f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.747434] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd70c763-5a1a-4bd7-9b8d-e218cf48061f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.754514] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f47edcb-3b66-4f34-8225-5e10f0792b6a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.769162] env[61006]: DEBUG nova.compute.provider_tree [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.134447] env[61006]: DEBUG nova.network.neutron [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.190143] env[61006]: DEBUG nova.network.neutron [-] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.224952] env[61006]: DEBUG nova.network.neutron [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.272538] env[61006]: DEBUG nova.scheduler.client.report [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.508501] env[61006]: DEBUG nova.network.neutron [-] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.693056] env[61006]: INFO nova.compute.manager [-] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Took 1.03 seconds to deallocate network for instance. [ 720.695216] env[61006]: DEBUG nova.compute.claims [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 720.695392] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.727122] env[61006]: DEBUG oslo_concurrency.lockutils [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] Releasing lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.727413] env[61006]: DEBUG nova.compute.manager [req-baec2958-26d9-4b6f-8a0c-807e85d76bc4 req-6b1a9f51-d5e6-44f1-8e91-cd8c6a33ca69 service nova] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Received event network-vif-deleted-374ad675-7064-4009-afd1-7727266d7e86 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.777486] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.778958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.487s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.011039] env[61006]: INFO nova.compute.manager [-] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Took 1.54 seconds to deallocate network for instance. [ 721.013469] env[61006]: DEBUG nova.compute.claims [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.013648] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.288050] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "bbb5f081-c791-4541-a5fb-723e166502c9" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.288050] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "bbb5f081-c791-4541-a5fb-723e166502c9" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.616028] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a2fe0c-db42-4eb1-af6b-6ca28ca9c172 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.624313] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bef00f5-0ee9-4e2f-a9fa-3c06f5f519c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.654143] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f0926c-59ea-45d4-988e-39c11fa21a8c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.661166] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d87116-535a-4ca2-bcae-40b63a776e13 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.673690] env[61006]: DEBUG nova.compute.provider_tree [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.790084] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "bbb5f081-c791-4541-a5fb-723e166502c9" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.790681] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.177198] env[61006]: DEBUG nova.scheduler.client.report [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.296088] env[61006]: DEBUG nova.compute.utils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.297481] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.297675] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 722.348363] env[61006]: DEBUG nova.policy [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70dd61a70249493cae18e9005b4e5b71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e1dc0db639c4db2a7aa10a3712cdddd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.650845] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Successfully created port: 4b7ae043-10d0-4455-a0fe-567d75e9232d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.682217] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.683079] env[61006]: ERROR nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Traceback (most recent call last): [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.driver.spawn(context, instance, image_meta, [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] vm_ref = self.build_virtual_machine(instance, [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.683079] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] for vif in network_info: [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self._sync_wrapper(fn, *args, **kwargs) [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.wait() [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self[:] = self._gt.wait() [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self._exit_event.wait() [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] result = hub.switch() [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 722.683426] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return self.greenlet.switch() [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] result = function(*args, **kwargs) [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] return func(*args, **kwargs) [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise e [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] nwinfo = self.network_api.allocate_for_instance( [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] created_port_ids = self._update_ports_for_instance( [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] with excutils.save_and_reraise_exception(): [ 722.683770] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] self.force_reraise() [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise self.value [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] updated_port = self._update_port( [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] _ensure_no_port_binding_failure(port) [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] raise exception.PortBindingFailed(port_id=port['id']) [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] nova.exception.PortBindingFailed: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. [ 722.684173] env[61006]: ERROR nova.compute.manager [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] [ 722.684471] env[61006]: DEBUG nova.compute.utils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.685388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.016s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.688291] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Build of instance c0b97c94-a055-4ce3-b961-1cdb6a5955a3 was re-scheduled: Binding failed for port 6fbf4c93-f06e-4138-ab07-95549266ba2e, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.688873] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.689215] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.689470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquired lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.689708] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 722.802831] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.217894] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.354103] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.504942] env[61006]: DEBUG nova.compute.manager [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Received event network-changed-4b7ae043-10d0-4455-a0fe-567d75e9232d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.505123] env[61006]: DEBUG nova.compute.manager [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Refreshing instance network info cache due to event network-changed-4b7ae043-10d0-4455-a0fe-567d75e9232d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 723.505365] env[61006]: DEBUG oslo_concurrency.lockutils [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] Acquiring lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.505525] env[61006]: DEBUG oslo_concurrency.lockutils [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] Acquired lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.505698] env[61006]: DEBUG nova.network.neutron [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Refreshing network info cache for port 4b7ae043-10d0-4455-a0fe-567d75e9232d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.628814] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0e3464-621f-4704-9c11-b1d1b6585378 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.636764] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4e178f-0aa1-4727-9e9f-f6c4333845a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.670537] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c11385-8b49-460f-a9f5-4c9b46a4ccd3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.680018] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374aea7f-e9c6-4872-a41f-7e802f2dc91c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.697024] env[61006]: DEBUG nova.compute.provider_tree [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.814959] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.851177] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.851177] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.851177] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.851177] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.851668] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.851668] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.851668] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.851668] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.851668] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.851934] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.851934] env[61006]: DEBUG nova.virt.hardware [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.851934] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9329557e-f78c-4d62-a343-f72baeeb63eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.859328] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Releasing lock "refresh_cache-c0b97c94-a055-4ce3-b961-1cdb6a5955a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.862020] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 723.862020] env[61006]: DEBUG nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.862020] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.862852] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34afd66c-0f3a-4863-846d-a4d18879a08e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.885374] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.887902] env[61006]: ERROR nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 723.887902] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.887902] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.887902] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.887902] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.887902] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.887902] env[61006]: ERROR nova.compute.manager raise self.value [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.887902] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.887902] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.887902] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.888474] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.888474] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.888474] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 723.888474] env[61006]: ERROR nova.compute.manager [ 723.888474] env[61006]: Traceback (most recent call last): [ 723.888474] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.888474] env[61006]: listener.cb(fileno) [ 723.888474] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.888474] env[61006]: result = function(*args, **kwargs) [ 723.888474] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.888474] env[61006]: return func(*args, **kwargs) [ 723.888474] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.888474] env[61006]: raise e [ 723.888474] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.888474] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 723.888474] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.888474] env[61006]: created_port_ids = self._update_ports_for_instance( [ 723.888474] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.888474] env[61006]: with excutils.save_and_reraise_exception(): [ 723.888474] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.888474] env[61006]: self.force_reraise() [ 723.888474] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.888474] env[61006]: raise self.value [ 723.888474] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.888474] env[61006]: updated_port = self._update_port( [ 723.888474] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.888474] env[61006]: _ensure_no_port_binding_failure(port) [ 723.888474] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.888474] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.889380] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 723.889380] env[61006]: Removing descriptor: 17 [ 723.889533] env[61006]: ERROR nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Traceback (most recent call last): [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] yield resources [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.driver.spawn(context, instance, image_meta, [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] vm_ref = self.build_virtual_machine(instance, [ 723.889533] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] for vif in network_info: [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self._sync_wrapper(fn, *args, **kwargs) [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.wait() [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self[:] = self._gt.wait() [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self._exit_event.wait() [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.889911] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] result = hub.switch() [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self.greenlet.switch() [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] result = function(*args, **kwargs) [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return func(*args, **kwargs) [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise e [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] nwinfo = self.network_api.allocate_for_instance( [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] created_port_ids = self._update_ports_for_instance( [ 723.890274] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] with excutils.save_and_reraise_exception(): [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.force_reraise() [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise self.value [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] updated_port = self._update_port( [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] _ensure_no_port_binding_failure(port) [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise exception.PortBindingFailed(port_id=port['id']) [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 723.890666] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] [ 723.891048] env[61006]: INFO nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Terminating instance [ 723.894224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.023338] env[61006]: DEBUG nova.network.neutron [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.106153] env[61006]: DEBUG nova.network.neutron [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.201089] env[61006]: DEBUG nova.scheduler.client.report [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.388482] env[61006]: DEBUG nova.network.neutron [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.608877] env[61006]: DEBUG oslo_concurrency.lockutils [req-3909adb9-3b0b-4f64-8c1d-daa354a6eba1 req-304e381a-d17d-4b77-81aa-7c121c8a1d06 service nova] Releasing lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.609319] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquired lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.609501] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.706290] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.706946] env[61006]: ERROR nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] Traceback (most recent call last): [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.driver.spawn(context, instance, image_meta, [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] vm_ref = self.build_virtual_machine(instance, [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.706946] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] for vif in network_info: [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return self._sync_wrapper(fn, *args, **kwargs) [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.wait() [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self[:] = self._gt.wait() [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return self._exit_event.wait() [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] current.throw(*self._exc) [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.707469] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] result = function(*args, **kwargs) [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] return func(*args, **kwargs) [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise e [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] nwinfo = self.network_api.allocate_for_instance( [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] created_port_ids = self._update_ports_for_instance( [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] with excutils.save_and_reraise_exception(): [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] self.force_reraise() [ 724.707868] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise self.value [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] updated_port = self._update_port( [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] _ensure_no_port_binding_failure(port) [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] raise exception.PortBindingFailed(port_id=port['id']) [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] nova.exception.PortBindingFailed: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. [ 724.708283] env[61006]: ERROR nova.compute.manager [instance: f7325d28-ed07-4dab-b440-2daf766882de] [ 724.708283] env[61006]: DEBUG nova.compute.utils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.708910] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.847s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.711811] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Build of instance f7325d28-ed07-4dab-b440-2daf766882de was re-scheduled: Binding failed for port a5b9458d-b5b2-4340-b44b-a237ece911cc, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.712222] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.712441] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquiring lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.712582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Acquired lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.712737] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.891691] env[61006]: INFO nova.compute.manager [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: c0b97c94-a055-4ce3-b961-1cdb6a5955a3] Took 1.03 seconds to deallocate network for instance. [ 725.128394] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.223713] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.238785] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.316919] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.551731] env[61006]: DEBUG nova.compute.manager [req-960a8c11-770e-4cec-bed8-c7fe946a0c06 req-b40e7c89-f239-450e-8b2c-888c22f8b319 service nova] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Received event network-vif-deleted-4b7ae043-10d0-4455-a0fe-567d75e9232d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.581507] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86d2a1c-f508-4769-81db-b079d8d4f506 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.588766] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b00e68-6f4e-40e8-b0bc-c4be010eac36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.617716] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b5b3ef-c5f3-4519-8670-a5b0f69cd83f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.624924] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a55c96-0dcc-498b-b68e-07964a63c7e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.638967] env[61006]: DEBUG nova.compute.provider_tree [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.728024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Releasing lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.728024] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.728024] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.728024] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-559f271f-0572-4f2d-bcec-f3dee55aae6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.736422] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa8e89a-f2e2-478c-bb93-3a2b2099c66c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.756776] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0211d621-ba0e-480b-8abd-58df829a1e39 could not be found. [ 725.757066] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.757285] env[61006]: INFO nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Took 0.03 seconds to destroy the instance on the hypervisor. [ 725.757555] env[61006]: DEBUG oslo.service.loopingcall [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.757830] env[61006]: DEBUG nova.compute.manager [-] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.757948] env[61006]: DEBUG nova.network.neutron [-] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.772397] env[61006]: DEBUG nova.network.neutron [-] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.819568] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Releasing lock "refresh_cache-f7325d28-ed07-4dab-b440-2daf766882de" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.819800] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.819974] env[61006]: DEBUG nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.820151] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.833376] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.918766] env[61006]: INFO nova.scheduler.client.report [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Deleted allocations for instance c0b97c94-a055-4ce3-b961-1cdb6a5955a3 [ 726.142104] env[61006]: DEBUG nova.scheduler.client.report [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.274907] env[61006]: DEBUG nova.network.neutron [-] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.335939] env[61006]: DEBUG nova.network.neutron [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.426991] env[61006]: DEBUG oslo_concurrency.lockutils [None req-687e6bb5-485a-4451-988a-feb9bc129e8c tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "c0b97c94-a055-4ce3-b961-1cdb6a5955a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.108s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.647928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.648574] env[61006]: ERROR nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Traceback (most recent call last): [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.driver.spawn(context, instance, image_meta, [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] vm_ref = self.build_virtual_machine(instance, [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.648574] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] for vif in network_info: [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return self._sync_wrapper(fn, *args, **kwargs) [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.wait() [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self[:] = self._gt.wait() [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return self._exit_event.wait() [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] current.throw(*self._exc) [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.648957] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] result = function(*args, **kwargs) [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] return func(*args, **kwargs) [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise e [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] nwinfo = self.network_api.allocate_for_instance( [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] created_port_ids = self._update_ports_for_instance( [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] with excutils.save_and_reraise_exception(): [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] self.force_reraise() [ 726.649379] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise self.value [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] updated_port = self._update_port( [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] _ensure_no_port_binding_failure(port) [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] raise exception.PortBindingFailed(port_id=port['id']) [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] nova.exception.PortBindingFailed: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. [ 726.649801] env[61006]: ERROR nova.compute.manager [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] [ 726.649801] env[61006]: DEBUG nova.compute.utils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.650614] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.132s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.652083] env[61006]: INFO nova.compute.claims [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.655659] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Build of instance 5525a859-eb09-45a6-80d6-8a466976a7bf was re-scheduled: Binding failed for port fbadc913-f41e-4f04-b065-9052f88f9c4d, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.655659] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.655659] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquiring lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.655659] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Acquired lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.655920] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.777328] env[61006]: INFO nova.compute.manager [-] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Took 1.02 seconds to deallocate network for instance. [ 726.779542] env[61006]: DEBUG nova.compute.claims [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.779720] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.838757] env[61006]: INFO nova.compute.manager [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] [instance: f7325d28-ed07-4dab-b440-2daf766882de] Took 1.02 seconds to deallocate network for instance. [ 726.929263] env[61006]: DEBUG nova.compute.manager [None req-f120e63e-bd67-4579-a960-d4683612ed9d tempest-ServersListShow296Test-1888338655 tempest-ServersListShow296Test-1888338655-project-member] [instance: c1b4dfa5-fd20-4064-85ee-da85518a5a20] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.175063] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.268648] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.433290] env[61006]: DEBUG nova.compute.manager [None req-f120e63e-bd67-4579-a960-d4683612ed9d tempest-ServersListShow296Test-1888338655 tempest-ServersListShow296Test-1888338655-project-member] [instance: c1b4dfa5-fd20-4064-85ee-da85518a5a20] Instance disappeared before build. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 727.772250] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Releasing lock "refresh_cache-5525a859-eb09-45a6-80d6-8a466976a7bf" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.772520] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.772673] env[61006]: DEBUG nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.772854] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.789261] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.863086] env[61006]: INFO nova.scheduler.client.report [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Deleted allocations for instance f7325d28-ed07-4dab-b440-2daf766882de [ 727.945338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f120e63e-bd67-4579-a960-d4683612ed9d tempest-ServersListShow296Test-1888338655 tempest-ServersListShow296Test-1888338655-project-member] Lock "c1b4dfa5-fd20-4064-85ee-da85518a5a20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.268s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.998370] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180bea8a-c24c-4776-8935-77233752efe9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.006304] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14591450-cf3e-484b-a4ea-742abc9f032d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.036601] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05ad461-4737-4dd8-bbc7-0ecbccce5eb4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.043449] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d8d35e-d2b8-4ba5-b868-bf53a637f3dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.058385] env[61006]: DEBUG nova.compute.provider_tree [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.291730] env[61006]: DEBUG nova.network.neutron [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.373548] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdd6076c-d62c-4f5e-ac95-a59e63c27f37 tempest-ServersAdminTestJSON-1403799569 tempest-ServersAdminTestJSON-1403799569-project-member] Lock "f7325d28-ed07-4dab-b440-2daf766882de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.841s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.448889] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.561950] env[61006]: DEBUG nova.scheduler.client.report [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.795244] env[61006]: INFO nova.compute.manager [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] [instance: 5525a859-eb09-45a6-80d6-8a466976a7bf] Took 1.02 seconds to deallocate network for instance. [ 728.877016] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.971799] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.070022] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.070022] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.071555] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.613s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.079369] env[61006]: INFO nova.compute.claims [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.399453] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.588472] env[61006]: DEBUG nova.compute.utils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.594348] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.594598] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.666333] env[61006]: DEBUG nova.policy [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64a9de8a5666445da83e50ef2ecf6c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0bc90fb759e452984348521bfbcb100', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.841251] env[61006]: INFO nova.scheduler.client.report [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Deleted allocations for instance 5525a859-eb09-45a6-80d6-8a466976a7bf [ 730.021848] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Successfully created port: 676858a9-6259-41fc-8893-ba484cb3af7f {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.094726] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.352551] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1488047-3461-4063-b383-9f44efec9ea1 tempest-ServerActionsTestOtherA-779240327 tempest-ServerActionsTestOtherA-779240327-project-member] Lock "5525a859-eb09-45a6-80d6-8a466976a7bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.595s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.562848] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80af07f2-3a7a-4b0b-ab60-e66ae6308e2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.570947] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db5d30c-008a-4c11-bfd6-f61d2c0009b6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.601222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e52fd90-b577-42af-866f-9339c370965b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.612067] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777e363a-89af-4396-ad76-075c8fea2b5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.625097] env[61006]: DEBUG nova.compute.provider_tree [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.856205] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.965313] env[61006]: DEBUG nova.compute.manager [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Received event network-changed-676858a9-6259-41fc-8893-ba484cb3af7f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.965672] env[61006]: DEBUG nova.compute.manager [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Refreshing instance network info cache due to event network-changed-676858a9-6259-41fc-8893-ba484cb3af7f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.965959] env[61006]: DEBUG oslo_concurrency.lockutils [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] Acquiring lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.966178] env[61006]: DEBUG oslo_concurrency.lockutils [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] Acquired lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.966428] env[61006]: DEBUG nova.network.neutron [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Refreshing network info cache for port 676858a9-6259-41fc-8893-ba484cb3af7f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.109064] env[61006]: ERROR nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 731.109064] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.109064] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.109064] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.109064] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.109064] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.109064] env[61006]: ERROR nova.compute.manager raise self.value [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.109064] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.109064] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.109064] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.109663] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.109663] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.109663] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 731.109663] env[61006]: ERROR nova.compute.manager [ 731.109663] env[61006]: Traceback (most recent call last): [ 731.109663] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.109663] env[61006]: listener.cb(fileno) [ 731.109663] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.109663] env[61006]: result = function(*args, **kwargs) [ 731.109663] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.109663] env[61006]: return func(*args, **kwargs) [ 731.109663] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.109663] env[61006]: raise e [ 731.109663] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.109663] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 731.109663] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.109663] env[61006]: created_port_ids = self._update_ports_for_instance( [ 731.109663] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.109663] env[61006]: with excutils.save_and_reraise_exception(): [ 731.109663] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.109663] env[61006]: self.force_reraise() [ 731.109663] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.109663] env[61006]: raise self.value [ 731.109663] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.109663] env[61006]: updated_port = self._update_port( [ 731.109663] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.109663] env[61006]: _ensure_no_port_binding_failure(port) [ 731.109663] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.109663] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.110606] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 731.110606] env[61006]: Removing descriptor: 20 [ 731.110606] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.128165] env[61006]: DEBUG nova.scheduler.client.report [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.139978] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.140249] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.140602] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.140602] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.140740] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.140888] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.141093] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.141252] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.141410] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.141565] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.141771] env[61006]: DEBUG nova.virt.hardware [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.142630] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbe2e00-7d10-49b6-802f-73382d8d17a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.156178] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0bbfa3-d1ce-4dbd-bcbe-6de39026835b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.171108] env[61006]: ERROR nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Traceback (most recent call last): [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] yield resources [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.driver.spawn(context, instance, image_meta, [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] vm_ref = self.build_virtual_machine(instance, [ 731.171108] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] for vif in network_info: [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return self._sync_wrapper(fn, *args, **kwargs) [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.wait() [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self[:] = self._gt.wait() [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return self._exit_event.wait() [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.171465] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] current.throw(*self._exc) [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] result = function(*args, **kwargs) [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return func(*args, **kwargs) [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise e [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] nwinfo = self.network_api.allocate_for_instance( [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] created_port_ids = self._update_ports_for_instance( [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] with excutils.save_and_reraise_exception(): [ 731.171922] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.force_reraise() [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise self.value [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] updated_port = self._update_port( [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] _ensure_no_port_binding_failure(port) [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise exception.PortBindingFailed(port_id=port['id']) [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 731.172347] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] [ 731.173023] env[61006]: INFO nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Terminating instance [ 731.175122] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquiring lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.389773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.491733] env[61006]: DEBUG nova.network.neutron [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.581474] env[61006]: DEBUG nova.network.neutron [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.635382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.635903] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.638653] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.784s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.911319] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 731.911615] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.087306] env[61006]: DEBUG oslo_concurrency.lockutils [req-c594404c-f971-43a9-aa61-e5de1517a866 req-355a9db4-a9cd-45e5-b294-8cfaf5dd7ffd service nova] Releasing lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.088127] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquired lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.088214] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.143728] env[61006]: DEBUG nova.compute.utils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.148844] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.148844] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.225555] env[61006]: DEBUG nova.policy [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7153b47300d04215bdc9362b3d44280f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a2db22bce944f77b8e10d3febd06ff9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.424408] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.424408] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 732.424408] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 732.579124] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f838b5a-f503-4a16-85e6-c02d56c14cc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.586924] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415c9fba-22ad-4e6d-b517-86bf3fd0eddd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.623932] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.625564] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804005e5-77c7-4360-9d9b-381fc555195e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.633476] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123951a6-4750-4f62-8d7d-7debb7563f76 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.648348] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.651291] env[61006]: DEBUG nova.compute.provider_tree [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.685149] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Successfully created port: 4c12cb68-f994-49dd-b22f-5eede94b7027 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.731531] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.926360] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.926360] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.926360] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.926707] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.926707] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.926774] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.927282] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Didn't find any instances for network info cache update. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 732.927282] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.927282] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.928555] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.928555] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.928555] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.928555] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.928555] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 732.933385] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.157964] env[61006]: DEBUG nova.scheduler.client.report [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.180812] env[61006]: DEBUG nova.compute.manager [req-089d8c2a-c1fe-4bb6-b15c-6294026261e8 req-3ddcec32-a679-4cd8-a9fc-2898802b81a7 service nova] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Received event network-vif-deleted-676858a9-6259-41fc-8893-ba484cb3af7f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.234390] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Releasing lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.234893] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.235103] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.235704] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bcdecb0a-0bee-4d9d-8c59-68562366ab43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.245942] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1836db22-2094-402d-ad5f-f5f1cb2565d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.277620] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 232272f0-1bf5-436e-ae24-5efa391eef57 could not be found. [ 733.277620] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.277620] env[61006]: INFO nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.277620] env[61006]: DEBUG oslo.service.loopingcall [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.277620] env[61006]: DEBUG nova.compute.manager [-] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.277620] env[61006]: DEBUG nova.network.neutron [-] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.305796] env[61006]: DEBUG nova.network.neutron [-] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.436568] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.664345] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.665582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.027s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.666762] env[61006]: ERROR nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Traceback (most recent call last): [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.driver.spawn(context, instance, image_meta, [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] vm_ref = self.build_virtual_machine(instance, [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.666762] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] for vif in network_info: [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self._sync_wrapper(fn, *args, **kwargs) [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.wait() [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self[:] = self._gt.wait() [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self._exit_event.wait() [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] result = hub.switch() [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.667180] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return self.greenlet.switch() [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] result = function(*args, **kwargs) [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] return func(*args, **kwargs) [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise e [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] nwinfo = self.network_api.allocate_for_instance( [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] created_port_ids = self._update_ports_for_instance( [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] with excutils.save_and_reraise_exception(): [ 733.667581] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] self.force_reraise() [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise self.value [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] updated_port = self._update_port( [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] _ensure_no_port_binding_failure(port) [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] raise exception.PortBindingFailed(port_id=port['id']) [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] nova.exception.PortBindingFailed: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. [ 733.667953] env[61006]: ERROR nova.compute.manager [instance: d6892ecb-0fb8-44da-9362-8806c429626b] [ 733.670016] env[61006]: DEBUG nova.compute.utils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.670671] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.295s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.673647] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Build of instance d6892ecb-0fb8-44da-9362-8806c429626b was re-scheduled: Binding failed for port 31e7e06b-bbd4-4886-b54a-e373e0347595, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 733.674256] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 733.675769] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquiring lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.675769] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Acquired lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.675769] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.697690] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.697937] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.698110] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.698309] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.698632] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.698632] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.698778] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.698932] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.699117] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.699279] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.699447] env[61006]: DEBUG nova.virt.hardware [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.700499] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bf44f5-dac3-456b-9588-713eb62d9454 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.708895] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4989712-1693-46bd-88d4-2ab4ebe9c1c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.807315] env[61006]: DEBUG nova.network.neutron [-] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.947634] env[61006]: ERROR nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 733.947634] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.947634] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.947634] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.947634] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.947634] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.947634] env[61006]: ERROR nova.compute.manager raise self.value [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.947634] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.947634] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.947634] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.948362] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.948362] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.948362] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 733.948362] env[61006]: ERROR nova.compute.manager [ 733.948362] env[61006]: Traceback (most recent call last): [ 733.948362] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.948362] env[61006]: listener.cb(fileno) [ 733.948362] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.948362] env[61006]: result = function(*args, **kwargs) [ 733.948362] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.948362] env[61006]: return func(*args, **kwargs) [ 733.948362] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.948362] env[61006]: raise e [ 733.948362] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.948362] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 733.948362] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.948362] env[61006]: created_port_ids = self._update_ports_for_instance( [ 733.948362] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.948362] env[61006]: with excutils.save_and_reraise_exception(): [ 733.948362] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.948362] env[61006]: self.force_reraise() [ 733.948362] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.948362] env[61006]: raise self.value [ 733.948362] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.948362] env[61006]: updated_port = self._update_port( [ 733.948362] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.948362] env[61006]: _ensure_no_port_binding_failure(port) [ 733.948362] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.948362] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.949341] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 733.949341] env[61006]: Removing descriptor: 20 [ 733.949341] env[61006]: ERROR nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Traceback (most recent call last): [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] yield resources [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.driver.spawn(context, instance, image_meta, [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.949341] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] vm_ref = self.build_virtual_machine(instance, [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] for vif in network_info: [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self._sync_wrapper(fn, *args, **kwargs) [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.wait() [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self[:] = self._gt.wait() [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self._exit_event.wait() [ 733.949739] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] result = hub.switch() [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self.greenlet.switch() [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] result = function(*args, **kwargs) [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return func(*args, **kwargs) [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise e [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] nwinfo = self.network_api.allocate_for_instance( [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.950175] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] created_port_ids = self._update_ports_for_instance( [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] with excutils.save_and_reraise_exception(): [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.force_reraise() [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise self.value [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] updated_port = self._update_port( [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] _ensure_no_port_binding_failure(port) [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.950609] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise exception.PortBindingFailed(port_id=port['id']) [ 733.950987] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 733.950987] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] [ 733.950987] env[61006]: INFO nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Terminating instance [ 733.951243] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquiring lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.951398] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquired lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.951596] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.202678] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.310874] env[61006]: INFO nova.compute.manager [-] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Took 1.03 seconds to deallocate network for instance. [ 734.315601] env[61006]: DEBUG nova.compute.claims [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.315777] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.362347] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.481030] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.554093] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17323c2a-a644-4f50-ae55-83ffc822518e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.562015] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f26a841-618f-4e90-be57-10dd749e2500 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.565547] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.593701] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87f9151-a319-4101-87ed-f243d762534d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.601777] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9606877-4958-4745-a25f-81ee9970e2ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.620671] env[61006]: DEBUG nova.compute.provider_tree [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.864690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Releasing lock "refresh_cache-d6892ecb-0fb8-44da-9362-8806c429626b" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.864935] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 734.865130] env[61006]: DEBUG nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.865294] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.881013] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.068053] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Releasing lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.068496] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.068688] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.068998] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3869eea6-aa1a-4413-a61d-3e4c70f0f6e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.078294] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293bb5b9-d6bd-4f1a-97eb-114c561f581c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.098872] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c156407-4ac6-4e34-812d-d719f1cc33d0 could not be found. [ 735.099103] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.099284] env[61006]: INFO nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 735.099520] env[61006]: DEBUG oslo.service.loopingcall [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.099741] env[61006]: DEBUG nova.compute.manager [-] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.099837] env[61006]: DEBUG nova.network.neutron [-] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.117144] env[61006]: DEBUG nova.network.neutron [-] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.123494] env[61006]: DEBUG nova.scheduler.client.report [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.211549] env[61006]: DEBUG nova.compute.manager [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Received event network-changed-4c12cb68-f994-49dd-b22f-5eede94b7027 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.211822] env[61006]: DEBUG nova.compute.manager [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Refreshing instance network info cache due to event network-changed-4c12cb68-f994-49dd-b22f-5eede94b7027. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.212153] env[61006]: DEBUG oslo_concurrency.lockutils [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] Acquiring lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.212358] env[61006]: DEBUG oslo_concurrency.lockutils [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] Acquired lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.212579] env[61006]: DEBUG nova.network.neutron [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Refreshing network info cache for port 4c12cb68-f994-49dd-b22f-5eede94b7027 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.385446] env[61006]: DEBUG nova.network.neutron [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.620781] env[61006]: DEBUG nova.network.neutron [-] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.628659] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.629506] env[61006]: ERROR nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Traceback (most recent call last): [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.driver.spawn(context, instance, image_meta, [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] vm_ref = self.build_virtual_machine(instance, [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.629506] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] for vif in network_info: [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self._sync_wrapper(fn, *args, **kwargs) [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.wait() [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self[:] = self._gt.wait() [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self._exit_event.wait() [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] result = hub.switch() [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.630124] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return self.greenlet.switch() [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] result = function(*args, **kwargs) [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] return func(*args, **kwargs) [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise e [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] nwinfo = self.network_api.allocate_for_instance( [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] created_port_ids = self._update_ports_for_instance( [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] with excutils.save_and_reraise_exception(): [ 735.630845] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] self.force_reraise() [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise self.value [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] updated_port = self._update_port( [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] _ensure_no_port_binding_failure(port) [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] raise exception.PortBindingFailed(port_id=port['id']) [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] nova.exception.PortBindingFailed: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. [ 735.631552] env[61006]: ERROR nova.compute.manager [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] [ 735.632122] env[61006]: DEBUG nova.compute.utils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 735.632122] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Build of instance ea2a73fa-3ece-4c88-b117-3b9843a4f224 was re-scheduled: Binding failed for port 11aee232-a61b-434b-9bed-13c67e38a458, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 735.632632] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 735.632909] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquiring lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.633118] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Acquired lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.633334] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.634371] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.939s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.737571] env[61006]: DEBUG nova.network.neutron [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.879801] env[61006]: DEBUG nova.network.neutron [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.888803] env[61006]: INFO nova.compute.manager [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] [instance: d6892ecb-0fb8-44da-9362-8806c429626b] Took 1.02 seconds to deallocate network for instance. [ 736.123866] env[61006]: INFO nova.compute.manager [-] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Took 1.02 seconds to deallocate network for instance. [ 736.129412] env[61006]: DEBUG nova.compute.claims [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.129412] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.171324] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.322370] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.381981] env[61006]: DEBUG oslo_concurrency.lockutils [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] Releasing lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.382304] env[61006]: DEBUG nova.compute.manager [req-f288381c-77a7-4e36-8c41-dc9e15035cb2 req-01559556-5d69-40ac-8753-09934c722c9d service nova] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Received event network-vif-deleted-4c12cb68-f994-49dd-b22f-5eede94b7027 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.583080] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08478e0d-f362-43d8-9774-f3bf8cf16ce8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.594021] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39163f95-494c-445d-b3e1-7a98acd534b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.625963] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d17e64e-035f-4682-af12-9750b3b3b1b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.633286] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94cd76e-fd52-41c4-bd6c-7ea77acd3010 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.649823] env[61006]: DEBUG nova.compute.provider_tree [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.826574] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Releasing lock "refresh_cache-ea2a73fa-3ece-4c88-b117-3b9843a4f224" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.829155] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 736.829155] env[61006]: DEBUG nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.829155] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.844848] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.925165] env[61006]: INFO nova.scheduler.client.report [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Deleted allocations for instance d6892ecb-0fb8-44da-9362-8806c429626b [ 737.154120] env[61006]: DEBUG nova.scheduler.client.report [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.347721] env[61006]: DEBUG nova.network.neutron [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.434516] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ebc9ff2-6b03-44fa-9b5d-5747dde7153d tempest-ServerRescueTestJSON-891369644 tempest-ServerRescueTestJSON-891369644-project-member] Lock "d6892ecb-0fb8-44da-9362-8806c429626b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.172s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.659378] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.660020] env[61006]: ERROR nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Traceback (most recent call last): [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.driver.spawn(context, instance, image_meta, [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] vm_ref = self.build_virtual_machine(instance, [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.660020] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] for vif in network_info: [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self._sync_wrapper(fn, *args, **kwargs) [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.wait() [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self[:] = self._gt.wait() [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self._exit_event.wait() [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] result = hub.switch() [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.660493] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return self.greenlet.switch() [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] result = function(*args, **kwargs) [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] return func(*args, **kwargs) [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise e [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] nwinfo = self.network_api.allocate_for_instance( [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] created_port_ids = self._update_ports_for_instance( [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] with excutils.save_and_reraise_exception(): [ 737.660989] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] self.force_reraise() [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise self.value [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] updated_port = self._update_port( [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] _ensure_no_port_binding_failure(port) [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] raise exception.PortBindingFailed(port_id=port['id']) [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] nova.exception.PortBindingFailed: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. [ 737.661494] env[61006]: ERROR nova.compute.manager [instance: 2217133d-52fb-45c4-bb84-63dde0546747] [ 737.661926] env[61006]: DEBUG nova.compute.utils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.663141] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.648s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.666198] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Build of instance 2217133d-52fb-45c4-bb84-63dde0546747 was re-scheduled: Binding failed for port 374ad675-7064-4009-afd1-7727266d7e86, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.666651] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.666878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquiring lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.667079] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Acquired lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.667519] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.853230] env[61006]: INFO nova.compute.manager [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] [instance: ea2a73fa-3ece-4c88-b117-3b9843a4f224] Took 1.02 seconds to deallocate network for instance. [ 737.937558] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.193138] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.233710] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.234649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.307385] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.462602] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.534930] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6553b4cf-8e83-4652-908e-6a9e8f654749 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.547047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7de8c99-cf78-4573-a674-721139d3f54f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.579318] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05793600-69d1-442a-9b9f-f9285cc36566 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.587427] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e15472-4f2e-4678-8cfa-4609f1b3d66a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.604800] env[61006]: DEBUG nova.compute.provider_tree [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.810308] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Releasing lock "refresh_cache-2217133d-52fb-45c4-bb84-63dde0546747" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.810541] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.810792] env[61006]: DEBUG nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.810959] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.829389] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.885342] env[61006]: INFO nova.scheduler.client.report [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Deleted allocations for instance ea2a73fa-3ece-4c88-b117-3b9843a4f224 [ 739.107581] env[61006]: DEBUG nova.scheduler.client.report [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.335892] env[61006]: DEBUG nova.network.neutron [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.393388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cc9b4158-dc8b-4fe5-88b4-32c4dbf7925f tempest-ServerRescueTestJSONUnderV235-860936787 tempest-ServerRescueTestJSONUnderV235-860936787-project-member] Lock "ea2a73fa-3ece-4c88-b117-3b9843a4f224" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.774s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.613123] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.613781] env[61006]: ERROR nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Traceback (most recent call last): [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.driver.spawn(context, instance, image_meta, [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] vm_ref = self.build_virtual_machine(instance, [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.613781] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] for vif in network_info: [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self._sync_wrapper(fn, *args, **kwargs) [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.wait() [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self[:] = self._gt.wait() [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self._exit_event.wait() [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] result = hub.switch() [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.614117] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return self.greenlet.switch() [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] result = function(*args, **kwargs) [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] return func(*args, **kwargs) [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise e [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] nwinfo = self.network_api.allocate_for_instance( [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] created_port_ids = self._update_ports_for_instance( [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] with excutils.save_and_reraise_exception(): [ 739.614468] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] self.force_reraise() [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise self.value [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] updated_port = self._update_port( [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] _ensure_no_port_binding_failure(port) [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] raise exception.PortBindingFailed(port_id=port['id']) [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] nova.exception.PortBindingFailed: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. [ 739.614859] env[61006]: ERROR nova.compute.manager [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] [ 739.615172] env[61006]: DEBUG nova.compute.utils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.615936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.836s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.619586] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Build of instance 34f17493-0d3d-4dfd-a2ac-6adc9b388f95 was re-scheduled: Binding failed for port 3e55991e-7041-4159-b3b6-b2acb7def1f2, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 739.620407] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 739.620407] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.620407] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.621032] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.836229] env[61006]: INFO nova.compute.manager [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] [instance: 2217133d-52fb-45c4-bb84-63dde0546747] Took 1.03 seconds to deallocate network for instance. [ 739.895506] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.152855] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.248327] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.424442] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.490265] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd68a67c-6f0f-4d84-8ee9-2ff9e35ea26c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.499282] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778b92db-3936-42f5-9654-3de44a5ae35c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.535978] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603b7c8e-8b03-4ad9-87f1-7de35aefb830 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.545105] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c08fe5-f445-4526-8ac2-073de749d2db {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.559398] env[61006]: DEBUG nova.compute.provider_tree [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.752817] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Releasing lock "refresh_cache-34f17493-0d3d-4dfd-a2ac-6adc9b388f95" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.753096] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 740.753276] env[61006]: DEBUG nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.753457] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.773332] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.877611] env[61006]: INFO nova.scheduler.client.report [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Deleted allocations for instance 2217133d-52fb-45c4-bb84-63dde0546747 [ 741.062766] env[61006]: DEBUG nova.scheduler.client.report [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.275021] env[61006]: DEBUG nova.network.neutron [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.385388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e250a334-aac6-4409-8265-d1302e5ae24f tempest-ServersTestFqdnHostnames-920586761 tempest-ServersTestFqdnHostnames-920586761-project-member] Lock "2217133d-52fb-45c4-bb84-63dde0546747" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.780s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.567713] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.568379] env[61006]: ERROR nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Traceback (most recent call last): [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.driver.spawn(context, instance, image_meta, [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] vm_ref = self.build_virtual_machine(instance, [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.568379] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] for vif in network_info: [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self._sync_wrapper(fn, *args, **kwargs) [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.wait() [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self[:] = self._gt.wait() [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self._exit_event.wait() [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] result = hub.switch() [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.568781] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return self.greenlet.switch() [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] result = function(*args, **kwargs) [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] return func(*args, **kwargs) [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise e [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] nwinfo = self.network_api.allocate_for_instance( [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] created_port_ids = self._update_ports_for_instance( [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] with excutils.save_and_reraise_exception(): [ 741.569210] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] self.force_reraise() [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise self.value [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] updated_port = self._update_port( [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] _ensure_no_port_binding_failure(port) [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] raise exception.PortBindingFailed(port_id=port['id']) [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] nova.exception.PortBindingFailed: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. [ 741.569620] env[61006]: ERROR nova.compute.manager [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] [ 741.569969] env[61006]: DEBUG nova.compute.utils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.570381] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.599s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.571865] env[61006]: INFO nova.compute.claims [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.574386] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Build of instance 0211d621-ba0e-480b-8abd-58df829a1e39 was re-scheduled: Binding failed for port 4b7ae043-10d0-4455-a0fe-567d75e9232d, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 741.574807] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 741.575039] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquiring lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.575186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Acquired lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.575341] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.777647] env[61006]: INFO nova.compute.manager [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 34f17493-0d3d-4dfd-a2ac-6adc9b388f95] Took 1.02 seconds to deallocate network for instance. [ 741.888208] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.109944] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.305711] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.414622] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.809891] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Releasing lock "refresh_cache-0211d621-ba0e-480b-8abd-58df829a1e39" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.810182] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 742.811361] env[61006]: DEBUG nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.811361] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.814084] env[61006]: INFO nova.scheduler.client.report [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Deleted allocations for instance 34f17493-0d3d-4dfd-a2ac-6adc9b388f95 [ 742.856092] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.989721] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c646ee-04d2-4dfe-90bd-75b4ed5a2373 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.997900] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921f541b-27f9-425d-9125-6d93e66dda2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.027934] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2b2350-fdaf-4874-8191-ea0c32d526ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.035217] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471455f2-c913-4dfa-a357-61fe7389bca3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.048106] env[61006]: DEBUG nova.compute.provider_tree [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.323795] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ae15f7c7-5324-4646-a109-e4cf954fece0 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "34f17493-0d3d-4dfd-a2ac-6adc9b388f95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.838s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.358086] env[61006]: DEBUG nova.network.neutron [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.552565] env[61006]: DEBUG nova.scheduler.client.report [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.827074] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.866186] env[61006]: INFO nova.compute.manager [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] [instance: 0211d621-ba0e-480b-8abd-58df829a1e39] Took 1.05 seconds to deallocate network for instance. [ 744.060018] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.060018] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.061215] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.662s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.062947] env[61006]: INFO nova.compute.claims [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.354604] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.568477] env[61006]: DEBUG nova.compute.utils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.574009] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.574204] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.646777] env[61006]: DEBUG nova.policy [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cca6e2806aa45208ae618f6a78ccc0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe9eabeec9a941e68a9eae559e24ff4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.903918] env[61006]: INFO nova.scheduler.client.report [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Deleted allocations for instance 0211d621-ba0e-480b-8abd-58df829a1e39 [ 744.921560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "0111452e-1b4f-499c-932d-f31364d1a14c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.921560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.078663] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.101223] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Successfully created port: 236ab510-684c-453b-ad81-da8a8d22898d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.421216] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6940982e-0624-455a-bcfb-36e0e4a02d71 tempest-ServerGroupTestJSON-88292640 tempest-ServerGroupTestJSON-88292640-project-member] Lock "0211d621-ba0e-480b-8abd-58df829a1e39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.252s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.460293] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517b2e88-5043-4c91-8e0a-46168e9798d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.468324] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d554ca1b-ba38-406b-b628-1b6a1946ce26 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.501570] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbd2dac-d45e-4c49-b28a-603368c131a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.508926] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de847b8-33e8-418a-82a7-895d93ac76c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.522678] env[61006]: DEBUG nova.compute.provider_tree [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.670941] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.671046] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.924438] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.028816] env[61006]: DEBUG nova.scheduler.client.report [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.092202] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.124303] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.124541] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.124726] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.124866] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.125011] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.126181] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.126181] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.126181] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.126181] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.126181] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.126373] env[61006]: DEBUG nova.virt.hardware [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.126953] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51af455-68e8-4534-9164-00d505e446bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.138055] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adec294-24ed-42be-b593-0ba85fd50a99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.202411] env[61006]: DEBUG nova.compute.manager [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Received event network-changed-236ab510-684c-453b-ad81-da8a8d22898d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.202517] env[61006]: DEBUG nova.compute.manager [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Refreshing instance network info cache due to event network-changed-236ab510-684c-453b-ad81-da8a8d22898d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.202737] env[61006]: DEBUG oslo_concurrency.lockutils [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] Acquiring lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.202870] env[61006]: DEBUG oslo_concurrency.lockutils [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] Acquired lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.203177] env[61006]: DEBUG nova.network.neutron [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Refreshing network info cache for port 236ab510-684c-453b-ad81-da8a8d22898d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.411177] env[61006]: ERROR nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 746.411177] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.411177] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.411177] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.411177] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.411177] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.411177] env[61006]: ERROR nova.compute.manager raise self.value [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.411177] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.411177] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.411177] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.411662] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.411662] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.411662] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 746.411662] env[61006]: ERROR nova.compute.manager [ 746.411662] env[61006]: Traceback (most recent call last): [ 746.411662] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.411662] env[61006]: listener.cb(fileno) [ 746.411662] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.411662] env[61006]: result = function(*args, **kwargs) [ 746.411662] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.411662] env[61006]: return func(*args, **kwargs) [ 746.411662] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.411662] env[61006]: raise e [ 746.411662] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.411662] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 746.411662] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.411662] env[61006]: created_port_ids = self._update_ports_for_instance( [ 746.411662] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.411662] env[61006]: with excutils.save_and_reraise_exception(): [ 746.411662] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.411662] env[61006]: self.force_reraise() [ 746.411662] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.411662] env[61006]: raise self.value [ 746.411662] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.411662] env[61006]: updated_port = self._update_port( [ 746.411662] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.411662] env[61006]: _ensure_no_port_binding_failure(port) [ 746.411662] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.411662] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.412489] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 746.412489] env[61006]: Removing descriptor: 20 [ 746.412489] env[61006]: ERROR nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Traceback (most recent call last): [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] yield resources [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.driver.spawn(context, instance, image_meta, [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.412489] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] vm_ref = self.build_virtual_machine(instance, [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] for vif in network_info: [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self._sync_wrapper(fn, *args, **kwargs) [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.wait() [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self[:] = self._gt.wait() [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self._exit_event.wait() [ 746.412943] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] result = hub.switch() [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self.greenlet.switch() [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] result = function(*args, **kwargs) [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return func(*args, **kwargs) [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise e [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] nwinfo = self.network_api.allocate_for_instance( [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.413360] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] created_port_ids = self._update_ports_for_instance( [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] with excutils.save_and_reraise_exception(): [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.force_reraise() [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise self.value [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] updated_port = self._update_port( [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] _ensure_no_port_binding_failure(port) [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.413728] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise exception.PortBindingFailed(port_id=port['id']) [ 746.414109] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 746.414109] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] [ 746.414109] env[61006]: INFO nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Terminating instance [ 746.417616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.451335] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.538107] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.541019] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.543633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.152s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.550023] env[61006]: INFO nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.724333] env[61006]: DEBUG nova.network.neutron [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.832947] env[61006]: DEBUG nova.network.neutron [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.053710] env[61006]: DEBUG nova.compute.utils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.057784] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.057784] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.111055] env[61006]: DEBUG nova.policy [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '376291781f334b38b700e9921f188105', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'becb5c16e41c422b87349bb527a0b79a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.331436] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.331742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.333191] env[61006]: DEBUG oslo_concurrency.lockutils [req-7aa5ee72-8ac8-4df5-9d15-3f45f2baa505 req-3a2bef35-08da-45e2-9490-84b452fc1858 service nova] Releasing lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.333731] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.333906] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.383074] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Successfully created port: b572595f-70ee-4065-87bc-5d4eb2849b52 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.561303] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.856353] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.991468] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375a6969-b660-4791-a5d7-d43caed6ce6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.999517] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab30fb15-e13d-44a6-92ee-2496e60d0b03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.033368] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2aa1139-fe37-4c48-b71d-773952de2967 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.041083] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45747d30-9851-49fd-a896-7d1ba8cf7e36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.047262] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.056348] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.234789] env[61006]: DEBUG nova.compute.manager [req-acca124b-0af4-44b0-9065-e276feb4c51c req-c1b61f55-5137-4628-8512-a69d3e3ef8dd service nova] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Received event network-vif-deleted-236ab510-684c-453b-ad81-da8a8d22898d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.548869] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.549327] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.549520] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.549934] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e4ef38d-0fd6-423f-be65-624356a403c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.558844] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4217f3-42d8-41ec-a358-026c4a717e4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.569919] env[61006]: ERROR nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 748.569919] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.569919] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.569919] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.569919] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.569919] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.569919] env[61006]: ERROR nova.compute.manager raise self.value [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.569919] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 748.569919] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.569919] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 748.570477] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.570477] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 748.570477] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 748.570477] env[61006]: ERROR nova.compute.manager [ 748.570477] env[61006]: Traceback (most recent call last): [ 748.570477] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 748.570477] env[61006]: listener.cb(fileno) [ 748.570477] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.570477] env[61006]: result = function(*args, **kwargs) [ 748.570477] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.570477] env[61006]: return func(*args, **kwargs) [ 748.570477] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.570477] env[61006]: raise e [ 748.570477] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.570477] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 748.570477] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.570477] env[61006]: created_port_ids = self._update_ports_for_instance( [ 748.570477] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.570477] env[61006]: with excutils.save_and_reraise_exception(): [ 748.570477] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.570477] env[61006]: self.force_reraise() [ 748.570477] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.570477] env[61006]: raise self.value [ 748.570477] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.570477] env[61006]: updated_port = self._update_port( [ 748.570477] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.570477] env[61006]: _ensure_no_port_binding_failure(port) [ 748.570477] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.570477] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 748.571453] env[61006]: nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 748.571453] env[61006]: Removing descriptor: 20 [ 748.571453] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.574952] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.588118] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd415afa-1e5f-42ea-b288-4432f15a5503 could not be found. [ 748.588315] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.588498] env[61006]: INFO nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Took 0.04 seconds to destroy the instance on the hypervisor. [ 748.588839] env[61006]: DEBUG oslo.service.loopingcall [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.589648] env[61006]: DEBUG nova.compute.manager [-] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.589747] env[61006]: DEBUG nova.network.neutron [-] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.602371] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.602591] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.602766] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.602965] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.603123] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.603266] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.603459] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.603613] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.603774] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.603933] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.604112] env[61006]: DEBUG nova.virt.hardware [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.604918] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bbd100-d6a6-4b15-97b0-f82dc7362bcb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.608169] env[61006]: DEBUG nova.network.neutron [-] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.614931] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c025b55-ad11-4378-bd40-c6987240fcc4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.630102] env[61006]: ERROR nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Traceback (most recent call last): [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] yield resources [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.driver.spawn(context, instance, image_meta, [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] vm_ref = self.build_virtual_machine(instance, [ 748.630102] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] for vif in network_info: [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return self._sync_wrapper(fn, *args, **kwargs) [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.wait() [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self[:] = self._gt.wait() [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return self._exit_event.wait() [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.630472] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] current.throw(*self._exc) [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] result = function(*args, **kwargs) [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return func(*args, **kwargs) [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise e [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] nwinfo = self.network_api.allocate_for_instance( [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] created_port_ids = self._update_ports_for_instance( [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] with excutils.save_and_reraise_exception(): [ 748.630977] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.force_reraise() [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise self.value [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] updated_port = self._update_port( [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] _ensure_no_port_binding_failure(port) [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise exception.PortBindingFailed(port_id=port['id']) [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 748.631381] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] [ 748.631381] env[61006]: INFO nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Terminating instance [ 748.632384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.632578] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.632778] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.078130] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.078677] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.081166] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.645s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.081338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.082196] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 749.082196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.766s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.085010] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af8ba5a-4dd1-4a88-aaf7-a3849dfc4007 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.094770] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087628d6-68ef-42c9-b14e-5e42ecdad1e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.108387] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093bc3d1-5ba2-438e-b043-71cab0202cdb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.110941] env[61006]: DEBUG nova.network.neutron [-] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.116377] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b2f150-1e17-4c39-ae15-5d3cb29af142 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.148391] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181450MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 749.148559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.165018] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.234593] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.585927] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.587308] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.587481] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 749.615219] env[61006]: INFO nova.compute.manager [-] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Took 1.02 seconds to deallocate network for instance. [ 749.617132] env[61006]: DEBUG nova.compute.claims [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.617132] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.658148] env[61006]: DEBUG nova.policy [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f28f12c38e642b6839c985e43edd320', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9206b1330c4e4928b1b3af4e5c57e541', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.736868] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.737356] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 749.737578] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.738289] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a619634-9090-483d-a800-dcf436fab6cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.750574] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdcb79e-ef69-4e4a-97a7-909c40eaa867 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.778070] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a38192e-4fb8-4767-b1d9-3d8b59a76925 could not be found. [ 749.778345] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.778530] env[61006]: INFO nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Took 0.04 seconds to destroy the instance on the hypervisor. [ 749.778775] env[61006]: DEBUG oslo.service.loopingcall [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.781272] env[61006]: DEBUG nova.compute.manager [-] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.781386] env[61006]: DEBUG nova.network.neutron [-] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.802732] env[61006]: DEBUG nova.network.neutron [-] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.960870] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Successfully created port: ce883c90-2640-4e48-be26-5ff7136fda9e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.996748] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c605c1-35ff-4f79-8c10-864cddce2728 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.005040] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6780a7-0ddb-472b-9dca-72ebfd50206d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.036336] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b8200e-0458-4b8c-8b44-100125d12ea3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.043936] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecd7921-551a-4570-9662-f3fe88172a63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.058450] env[61006]: DEBUG nova.compute.provider_tree [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.094242] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.275704] env[61006]: DEBUG nova.compute.manager [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Received event network-changed-b572595f-70ee-4065-87bc-5d4eb2849b52 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.275936] env[61006]: DEBUG nova.compute.manager [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Refreshing instance network info cache due to event network-changed-b572595f-70ee-4065-87bc-5d4eb2849b52. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.276095] env[61006]: DEBUG oslo_concurrency.lockutils [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] Acquiring lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.276297] env[61006]: DEBUG oslo_concurrency.lockutils [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] Acquired lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.276378] env[61006]: DEBUG nova.network.neutron [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Refreshing network info cache for port b572595f-70ee-4065-87bc-5d4eb2849b52 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.305923] env[61006]: DEBUG nova.network.neutron [-] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.561178] env[61006]: DEBUG nova.scheduler.client.report [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.797490] env[61006]: DEBUG nova.network.neutron [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.812761] env[61006]: INFO nova.compute.manager [-] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Took 1.03 seconds to deallocate network for instance. [ 750.815069] env[61006]: DEBUG nova.compute.claims [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.815291] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.916786] env[61006]: DEBUG nova.network.neutron [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.962791] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 750.962791] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.962791] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.962791] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.962791] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.962791] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.962791] env[61006]: ERROR nova.compute.manager raise self.value [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.962791] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.962791] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.962791] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.963484] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.963484] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.963484] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 750.963484] env[61006]: ERROR nova.compute.manager [ 750.963484] env[61006]: Traceback (most recent call last): [ 750.963484] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.963484] env[61006]: listener.cb(fileno) [ 750.963484] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.963484] env[61006]: result = function(*args, **kwargs) [ 750.963484] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.963484] env[61006]: return func(*args, **kwargs) [ 750.963484] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.963484] env[61006]: raise e [ 750.963484] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.963484] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 750.963484] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.963484] env[61006]: created_port_ids = self._update_ports_for_instance( [ 750.963484] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.963484] env[61006]: with excutils.save_and_reraise_exception(): [ 750.963484] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.963484] env[61006]: self.force_reraise() [ 750.963484] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.963484] env[61006]: raise self.value [ 750.963484] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.963484] env[61006]: updated_port = self._update_port( [ 750.963484] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.963484] env[61006]: _ensure_no_port_binding_failure(port) [ 750.963484] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.963484] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.964441] env[61006]: nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 750.964441] env[61006]: Removing descriptor: 20 [ 751.066290] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.066961] env[61006]: ERROR nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Traceback (most recent call last): [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.driver.spawn(context, instance, image_meta, [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] vm_ref = self.build_virtual_machine(instance, [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.066961] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] for vif in network_info: [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return self._sync_wrapper(fn, *args, **kwargs) [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.wait() [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self[:] = self._gt.wait() [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return self._exit_event.wait() [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] current.throw(*self._exc) [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.067319] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] result = function(*args, **kwargs) [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] return func(*args, **kwargs) [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise e [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] nwinfo = self.network_api.allocate_for_instance( [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] created_port_ids = self._update_ports_for_instance( [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] with excutils.save_and_reraise_exception(): [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] self.force_reraise() [ 751.067673] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise self.value [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] updated_port = self._update_port( [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] _ensure_no_port_binding_failure(port) [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] raise exception.PortBindingFailed(port_id=port['id']) [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] nova.exception.PortBindingFailed: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. [ 751.068043] env[61006]: ERROR nova.compute.manager [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] [ 751.068043] env[61006]: DEBUG nova.compute.utils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.068858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.941s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.072229] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Build of instance 232272f0-1bf5-436e-ae24-5efa391eef57 was re-scheduled: Binding failed for port 676858a9-6259-41fc-8893-ba484cb3af7f, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.072749] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.072944] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquiring lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.073100] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Acquired lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.073263] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.103161] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.127376] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.127620] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.127771] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.127949] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.128099] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.128245] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.128441] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.128591] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.128746] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.128898] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.129079] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.129932] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883eb907-8af1-4da7-8e9d-4b7780409e28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.137857] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2771092-23a3-44cc-9f9c-cd1166ac1d21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.152358] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Traceback (most recent call last): [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] yield resources [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.driver.spawn(context, instance, image_meta, [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] vm_ref = self.build_virtual_machine(instance, [ 751.152358] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] for vif in network_info: [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return self._sync_wrapper(fn, *args, **kwargs) [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.wait() [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self[:] = self._gt.wait() [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return self._exit_event.wait() [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 751.152776] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] current.throw(*self._exc) [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] result = function(*args, **kwargs) [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return func(*args, **kwargs) [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise e [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] nwinfo = self.network_api.allocate_for_instance( [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] created_port_ids = self._update_ports_for_instance( [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] with excutils.save_and_reraise_exception(): [ 751.153176] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.force_reraise() [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise self.value [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] updated_port = self._update_port( [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] _ensure_no_port_binding_failure(port) [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise exception.PortBindingFailed(port_id=port['id']) [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 751.153607] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] [ 751.153607] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Terminating instance [ 751.153971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.154040] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.154201] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.420026] env[61006]: DEBUG oslo_concurrency.lockutils [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] Releasing lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.420026] env[61006]: DEBUG nova.compute.manager [req-61e74eb1-9434-49d1-8d65-e467f1719b9b req-9d182e4e-3ffa-4639-a4e3-99b88eece8df service nova] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Received event network-vif-deleted-b572595f-70ee-4065-87bc-5d4eb2849b52 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.595038] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.651578] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.672446] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.756705] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.906949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d6ee76-6c88-47c8-84a8-0372663cbf7a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.914774] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650b52d5-502d-4157-9e40-bae32b47805f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.946113] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e2cc78-3906-43b7-a170-1dc27af07ae5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.953749] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b5fed7-a1c1-4fe1-8ef3-a3b1058aa8ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.966443] env[61006]: DEBUG nova.compute.provider_tree [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.154492] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Releasing lock "refresh_cache-232272f0-1bf5-436e-ae24-5efa391eef57" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.154780] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.154881] env[61006]: DEBUG nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.155058] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.176528] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.262052] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.262052] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.262208] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.262447] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fc4b6d3-6a5c-4dfe-9de8-84d40fd0be4d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.271913] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e8442d-8173-42bc-99c8-16347768f5c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.293023] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b8fb9290-755f-40b8-af3d-c17407d0c846 could not be found. [ 752.293297] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.293486] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.293726] env[61006]: DEBUG oslo.service.loopingcall [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.293946] env[61006]: DEBUG nova.compute.manager [-] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.294060] env[61006]: DEBUG nova.network.neutron [-] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.300498] env[61006]: DEBUG nova.compute.manager [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Received event network-changed-ce883c90-2640-4e48-be26-5ff7136fda9e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.300707] env[61006]: DEBUG nova.compute.manager [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Refreshing instance network info cache due to event network-changed-ce883c90-2640-4e48-be26-5ff7136fda9e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 752.300884] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] Acquiring lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.301015] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] Acquired lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.301170] env[61006]: DEBUG nova.network.neutron [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Refreshing network info cache for port ce883c90-2640-4e48-be26-5ff7136fda9e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 752.314703] env[61006]: DEBUG nova.network.neutron [-] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.469040] env[61006]: DEBUG nova.scheduler.client.report [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.678804] env[61006]: DEBUG nova.network.neutron [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.817179] env[61006]: DEBUG nova.network.neutron [-] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.820428] env[61006]: DEBUG nova.network.neutron [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.901437] env[61006]: DEBUG nova.network.neutron [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.974904] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.975585] env[61006]: ERROR nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Traceback (most recent call last): [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.driver.spawn(context, instance, image_meta, [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] vm_ref = self.build_virtual_machine(instance, [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.975585] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] for vif in network_info: [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self._sync_wrapper(fn, *args, **kwargs) [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.wait() [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self[:] = self._gt.wait() [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self._exit_event.wait() [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] result = hub.switch() [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.975961] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return self.greenlet.switch() [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] result = function(*args, **kwargs) [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] return func(*args, **kwargs) [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise e [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] nwinfo = self.network_api.allocate_for_instance( [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] created_port_ids = self._update_ports_for_instance( [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] with excutils.save_and_reraise_exception(): [ 752.976395] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] self.force_reraise() [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise self.value [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] updated_port = self._update_port( [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] _ensure_no_port_binding_failure(port) [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] raise exception.PortBindingFailed(port_id=port['id']) [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] nova.exception.PortBindingFailed: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. [ 752.976805] env[61006]: ERROR nova.compute.manager [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] [ 752.977178] env[61006]: DEBUG nova.compute.utils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.977469] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.515s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.978971] env[61006]: INFO nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.982878] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Build of instance 6c156407-4ac6-4e34-812d-d719f1cc33d0 was re-scheduled: Binding failed for port 4c12cb68-f994-49dd-b22f-5eede94b7027, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.982878] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.982878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquiring lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.982878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Acquired lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.983130] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.181249] env[61006]: INFO nova.compute.manager [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] [instance: 232272f0-1bf5-436e-ae24-5efa391eef57] Took 1.03 seconds to deallocate network for instance. [ 753.320995] env[61006]: INFO nova.compute.manager [-] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Took 1.03 seconds to deallocate network for instance. [ 753.322985] env[61006]: DEBUG nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.323174] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.405860] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] Releasing lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.405989] env[61006]: DEBUG nova.compute.manager [req-da9bd512-f531-42ce-9856-425cefef7efa req-cce527bc-14ed-4b80-9fa6-76836fae63de service nova] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Received event network-vif-deleted-ce883c90-2640-4e48-be26-5ff7136fda9e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.503182] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.592541] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.096736] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Releasing lock "refresh_cache-6c156407-4ac6-4e34-812d-d719f1cc33d0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.096736] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.096736] env[61006]: DEBUG nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.096949] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.110562] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.207428] env[61006]: INFO nova.scheduler.client.report [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Deleted allocations for instance 232272f0-1bf5-436e-ae24-5efa391eef57 [ 754.300093] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48286f9-187c-491a-a48b-3fe3929c4e7e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.308237] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effa41cb-d7aa-4a65-b12d-53fd112adda6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.339641] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857551a4-577e-49f2-b6f2-dd998ea9c084 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.350895] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cceb18ae-df9f-4026-9956-fe249396bca3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.367159] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.612991] env[61006]: DEBUG nova.network.neutron [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.718563] env[61006]: DEBUG oslo_concurrency.lockutils [None req-65f3d049-3118-4fdc-bec0-ab81207892be tempest-ServerDiagnosticsTest-762660010 tempest-ServerDiagnosticsTest-762660010-project-member] Lock "232272f0-1bf5-436e-ae24-5efa391eef57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.171s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.871124] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.115839] env[61006]: INFO nova.compute.manager [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] [instance: 6c156407-4ac6-4e34-812d-d719f1cc33d0] Took 1.02 seconds to deallocate network for instance. [ 755.221244] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.379958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.380504] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.383105] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.959s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.384462] env[61006]: INFO nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.743708] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.889804] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.893549] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.893549] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.948873] env[61006]: DEBUG nova.policy [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f28f12c38e642b6839c985e43edd320', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9206b1330c4e4928b1b3af4e5c57e541', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.149036] env[61006]: INFO nova.scheduler.client.report [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Deleted allocations for instance 6c156407-4ac6-4e34-812d-d719f1cc33d0 [ 756.247134] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Successfully created port: 5b65a13b-f572-4137-80ce-2adecd295af8 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.394076] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.657934] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e483d70-4a9a-4843-9f48-7986751a7311 tempest-ServerActionsTestJSON-1074765716 tempest-ServerActionsTestJSON-1074765716-project-member] Lock "6c156407-4ac6-4e34-812d-d719f1cc33d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.117s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.768958] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171a630d-04ef-46f1-9b94-2bdc198b245a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.776832] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3d02d2-6c5c-4822-965f-cc46d7af5a5d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.811779] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87ba364-4422-4fae-a03e-17659348640c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.821172] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38af835-a428-411c-9c42-3909e890278f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.834594] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.058296] env[61006]: DEBUG nova.compute.manager [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Received event network-changed-5b65a13b-f572-4137-80ce-2adecd295af8 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.058296] env[61006]: DEBUG nova.compute.manager [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Refreshing instance network info cache due to event network-changed-5b65a13b-f572-4137-80ce-2adecd295af8. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.058503] env[61006]: DEBUG oslo_concurrency.lockutils [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] Acquiring lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.058653] env[61006]: DEBUG oslo_concurrency.lockutils [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] Acquired lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.058829] env[61006]: DEBUG nova.network.neutron [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Refreshing network info cache for port 5b65a13b-f572-4137-80ce-2adecd295af8 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.163462] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.338606] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.380999] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 757.380999] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.380999] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.380999] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.380999] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.380999] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.380999] env[61006]: ERROR nova.compute.manager raise self.value [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.380999] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.380999] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.380999] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.381605] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.381605] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.381605] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 757.381605] env[61006]: ERROR nova.compute.manager [ 757.381605] env[61006]: Traceback (most recent call last): [ 757.381605] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.381605] env[61006]: listener.cb(fileno) [ 757.381605] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.381605] env[61006]: result = function(*args, **kwargs) [ 757.381605] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.381605] env[61006]: return func(*args, **kwargs) [ 757.381605] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.381605] env[61006]: raise e [ 757.381605] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.381605] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 757.381605] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.381605] env[61006]: created_port_ids = self._update_ports_for_instance( [ 757.381605] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.381605] env[61006]: with excutils.save_and_reraise_exception(): [ 757.381605] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.381605] env[61006]: self.force_reraise() [ 757.381605] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.381605] env[61006]: raise self.value [ 757.381605] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.381605] env[61006]: updated_port = self._update_port( [ 757.381605] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.381605] env[61006]: _ensure_no_port_binding_failure(port) [ 757.381605] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.381605] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.382593] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 757.382593] env[61006]: Removing descriptor: 20 [ 757.410566] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.443017] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.443017] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.443017] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.443252] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.443252] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.443252] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.443252] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.443252] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.443500] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.443500] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.443500] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.443500] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59400662-4f5e-4e9d-b931-404e16c20bf8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.454242] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766a726a-86a1-4a02-b13d-1df8bd6b7468 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.470315] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Traceback (most recent call last): [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] yield resources [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.driver.spawn(context, instance, image_meta, [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] vm_ref = self.build_virtual_machine(instance, [ 757.470315] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] for vif in network_info: [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return self._sync_wrapper(fn, *args, **kwargs) [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.wait() [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self[:] = self._gt.wait() [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return self._exit_event.wait() [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.471073] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] current.throw(*self._exc) [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] result = function(*args, **kwargs) [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return func(*args, **kwargs) [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise e [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] nwinfo = self.network_api.allocate_for_instance( [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] created_port_ids = self._update_ports_for_instance( [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] with excutils.save_and_reraise_exception(): [ 757.472061] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.force_reraise() [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise self.value [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] updated_port = self._update_port( [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] _ensure_no_port_binding_failure(port) [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise exception.PortBindingFailed(port_id=port['id']) [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 757.472608] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] [ 757.472608] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Terminating instance [ 757.473679] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.598682] env[61006]: DEBUG nova.network.neutron [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.690090] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.757748] env[61006]: DEBUG nova.network.neutron [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.844207] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.844727] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.847662] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.433s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.849070] env[61006]: INFO nova.compute.claims [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.260333] env[61006]: DEBUG oslo_concurrency.lockutils [req-71262055-03d9-4dd5-8df2-5f73bbd7b3b0 req-d2e871b2-4f08-4621-bbdd-6030a7222940 service nova] Releasing lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.261030] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.261030] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.360394] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.362594] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.362779] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.477421] env[61006]: DEBUG nova.policy [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f28f12c38e642b6839c985e43edd320', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9206b1330c4e4928b1b3af4e5c57e541', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.816313] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.868127] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.045662] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.087117] env[61006]: DEBUG nova.compute.manager [req-9da5bffe-cc96-4cb2-96e9-15c8973432eb req-be4f1321-f8be-4b16-a75c-b9e1d5eaaa06 service nova] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Received event network-vif-deleted-5b65a13b-f572-4137-80ce-2adecd295af8 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.255063] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b42603-d858-4547-9498-855e218913ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.267515] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca293643-352d-400b-b3bc-760907af9fc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.301968] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66aaeabc-40d7-4f79-ab63-5cd4128dbb82 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.310111] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4eb795-1a26-4543-97d5-2509c183ee3a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.327156] env[61006]: DEBUG nova.compute.provider_tree [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.460607] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Successfully created port: c18dcc73-eae1-4b1f-b584-4f31b9d413f5 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.550455] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.550889] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.551092] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.551399] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c641fdb8-bfef-43e5-9857-cab4d861d0f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.560456] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e18db3c-58a9-43a8-8e7c-decae7029d2c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.582453] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a could not be found. [ 759.582741] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.582865] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 759.583139] env[61006]: DEBUG oslo.service.loopingcall [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.584248] env[61006]: DEBUG nova.compute.manager [-] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.584248] env[61006]: DEBUG nova.network.neutron [-] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.609615] env[61006]: DEBUG nova.network.neutron [-] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.834996] env[61006]: DEBUG nova.scheduler.client.report [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.882114] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.916021] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.916290] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.916464] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.916660] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.916804] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.916947] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.917250] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.917506] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.917782] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.917957] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.918283] env[61006]: DEBUG nova.virt.hardware [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.920532] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca24a6d-d3b6-4c12-88d0-392695ca6e3a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.929618] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632f0d91-fde8-4c78-89a1-0f93661e508f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.113510] env[61006]: DEBUG nova.network.neutron [-] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.341183] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.341676] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.345377] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.991s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.347378] env[61006]: INFO nova.compute.claims [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.616792] env[61006]: INFO nova.compute.manager [-] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Took 1.03 seconds to deallocate network for instance. [ 760.620800] env[61006]: DEBUG nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.620800] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.846921] env[61006]: DEBUG nova.compute.utils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.848298] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.848466] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.947768] env[61006]: DEBUG nova.policy [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '376291781f334b38b700e9921f188105', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'becb5c16e41c422b87349bb527a0b79a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.108500] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 761.108500] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.108500] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.108500] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.108500] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.108500] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.108500] env[61006]: ERROR nova.compute.manager raise self.value [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.108500] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.108500] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.108500] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.109095] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.109095] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.109095] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 761.109095] env[61006]: ERROR nova.compute.manager [ 761.109095] env[61006]: Traceback (most recent call last): [ 761.109095] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.109095] env[61006]: listener.cb(fileno) [ 761.109095] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.109095] env[61006]: result = function(*args, **kwargs) [ 761.109095] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.109095] env[61006]: return func(*args, **kwargs) [ 761.109095] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.109095] env[61006]: raise e [ 761.109095] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.109095] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 761.109095] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.109095] env[61006]: created_port_ids = self._update_ports_for_instance( [ 761.109095] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.109095] env[61006]: with excutils.save_and_reraise_exception(): [ 761.109095] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.109095] env[61006]: self.force_reraise() [ 761.109095] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.109095] env[61006]: raise self.value [ 761.109095] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.109095] env[61006]: updated_port = self._update_port( [ 761.109095] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.109095] env[61006]: _ensure_no_port_binding_failure(port) [ 761.109095] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.109095] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.110100] env[61006]: nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 761.110100] env[61006]: Removing descriptor: 20 [ 761.110100] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Traceback (most recent call last): [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] yield resources [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.driver.spawn(context, instance, image_meta, [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.110100] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] vm_ref = self.build_virtual_machine(instance, [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] for vif in network_info: [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self._sync_wrapper(fn, *args, **kwargs) [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.wait() [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self[:] = self._gt.wait() [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self._exit_event.wait() [ 761.110515] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] result = hub.switch() [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self.greenlet.switch() [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] result = function(*args, **kwargs) [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return func(*args, **kwargs) [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise e [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] nwinfo = self.network_api.allocate_for_instance( [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.110971] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] created_port_ids = self._update_ports_for_instance( [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] with excutils.save_and_reraise_exception(): [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.force_reraise() [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise self.value [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] updated_port = self._update_port( [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] _ensure_no_port_binding_failure(port) [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.111443] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise exception.PortBindingFailed(port_id=port['id']) [ 761.111836] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 761.111836] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] [ 761.111836] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Terminating instance [ 761.112210] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.112373] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.112536] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.340114] env[61006]: DEBUG nova.compute.manager [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Received event network-changed-c18dcc73-eae1-4b1f-b584-4f31b9d413f5 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.340114] env[61006]: DEBUG nova.compute.manager [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Refreshing instance network info cache due to event network-changed-c18dcc73-eae1-4b1f-b584-4f31b9d413f5. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.340114] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] Acquiring lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.353854] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.416074] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Successfully created port: f2a13097-5794-4ace-ac92-a25213089df1 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.649816] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.803938] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.838045] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.838290] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.934156] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4298d04-73d2-41a7-a0ee-bf1af3b614ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.943183] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832ccd15-31c6-4fed-9299-7897ac80ede0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.974211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9083ebfc-359e-470d-afe9-3d3ae3ccc633 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.982463] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aea46ab-9db8-45d6-bc15-704de5d35468 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.996487] env[61006]: DEBUG nova.compute.provider_tree [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.311019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.311019] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.311019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.311019] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] Acquired lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.311019] env[61006]: DEBUG nova.network.neutron [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Refreshing network info cache for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.312133] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17c09cd1-faa3-4112-8ab8-68b1c95aacd9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.322414] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e6dda0-4a04-4a71-abe3-57201d31714d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.348296] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ad1f43f-9e98-43b7-8076-20824ccc18ed could not be found. [ 762.348556] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.348733] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Took 0.04 seconds to destroy the instance on the hypervisor. [ 762.348976] env[61006]: DEBUG oslo.service.loopingcall [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.349214] env[61006]: DEBUG nova.compute.manager [-] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.349308] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.368356] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.373810] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.401725] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.402694] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.402694] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.402694] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.402694] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.402694] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.402863] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.402952] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.403246] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.403478] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.403704] env[61006]: DEBUG nova.virt.hardware [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.404737] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2a292d-960d-4a05-b6be-e8eac337fa19 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.413533] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b92478d-2bc2-4a58-90e7-59325754ed36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.499682] env[61006]: DEBUG nova.scheduler.client.report [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.546093] env[61006]: ERROR nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 762.546093] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.546093] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.546093] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.546093] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.546093] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.546093] env[61006]: ERROR nova.compute.manager raise self.value [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.546093] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 762.546093] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.546093] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 762.546445] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.546445] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 762.546445] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 762.546445] env[61006]: ERROR nova.compute.manager [ 762.546445] env[61006]: Traceback (most recent call last): [ 762.546445] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 762.546445] env[61006]: listener.cb(fileno) [ 762.546445] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.546445] env[61006]: result = function(*args, **kwargs) [ 762.546445] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.546445] env[61006]: return func(*args, **kwargs) [ 762.546445] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.546445] env[61006]: raise e [ 762.546445] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.546445] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 762.546445] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.546445] env[61006]: created_port_ids = self._update_ports_for_instance( [ 762.546445] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.546445] env[61006]: with excutils.save_and_reraise_exception(): [ 762.546445] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.546445] env[61006]: self.force_reraise() [ 762.546445] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.546445] env[61006]: raise self.value [ 762.546445] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.546445] env[61006]: updated_port = self._update_port( [ 762.546445] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.546445] env[61006]: _ensure_no_port_binding_failure(port) [ 762.546445] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.546445] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 762.547012] env[61006]: nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 762.547012] env[61006]: Removing descriptor: 17 [ 762.547012] env[61006]: ERROR nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Traceback (most recent call last): [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] yield resources [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.driver.spawn(context, instance, image_meta, [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.547012] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] vm_ref = self.build_virtual_machine(instance, [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] for vif in network_info: [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self._sync_wrapper(fn, *args, **kwargs) [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.wait() [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self[:] = self._gt.wait() [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self._exit_event.wait() [ 762.547356] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] result = hub.switch() [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self.greenlet.switch() [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] result = function(*args, **kwargs) [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return func(*args, **kwargs) [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise e [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] nwinfo = self.network_api.allocate_for_instance( [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.547619] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] created_port_ids = self._update_ports_for_instance( [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] with excutils.save_and_reraise_exception(): [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.force_reraise() [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise self.value [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] updated_port = self._update_port( [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] _ensure_no_port_binding_failure(port) [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.547893] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise exception.PortBindingFailed(port_id=port['id']) [ 762.548173] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 762.548173] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] [ 762.548173] env[61006]: INFO nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Terminating instance [ 762.550630] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.550630] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.550630] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.832247] env[61006]: DEBUG nova.network.neutron [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.879382] env[61006]: DEBUG nova.network.neutron [-] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.923166] env[61006]: DEBUG nova.network.neutron [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.007490] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.007490] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.010346] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.559s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.012872] env[61006]: INFO nova.compute.claims [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.071574] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.165186] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.382189] env[61006]: INFO nova.compute.manager [-] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Took 1.03 seconds to deallocate network for instance. [ 763.385860] env[61006]: DEBUG nova.compute.manager [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Received event network-changed-f2a13097-5794-4ace-ac92-a25213089df1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.386152] env[61006]: DEBUG nova.compute.manager [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Refreshing instance network info cache due to event network-changed-f2a13097-5794-4ace-ac92-a25213089df1. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 763.386436] env[61006]: DEBUG oslo_concurrency.lockutils [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] Acquiring lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.386890] env[61006]: DEBUG nova.compute.claims [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.387181] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.426862] env[61006]: DEBUG oslo_concurrency.lockutils [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] Releasing lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.427305] env[61006]: DEBUG nova.compute.manager [req-ef734ffa-555b-4a20-9dc9-5c9910d829ae req-d467f8ce-3c9b-4a8c-ae8e-5d83cb5b283a service nova] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Received event network-vif-deleted-c18dcc73-eae1-4b1f-b584-4f31b9d413f5 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.518643] env[61006]: DEBUG nova.compute.utils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.522288] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 763.522437] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.584937] env[61006]: DEBUG nova.policy [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8b0db2570144795b04809d82e643764', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ad8b37b118c4c8a8fde488ffdc44621', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.668989] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.669499] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 763.669693] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.670008] env[61006]: DEBUG oslo_concurrency.lockutils [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] Acquired lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.670189] env[61006]: DEBUG nova.network.neutron [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Refreshing network info cache for port f2a13097-5794-4ace-ac92-a25213089df1 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.671815] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e948e72b-5f8e-48f1-8cf9-55690ac5f13a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.682127] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc45e934-5dac-4c77-ad67-22c89767d212 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.703747] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e4a5a57e-f078-48ec-afe5-204c75139c15 could not be found. [ 763.703983] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.704179] env[61006]: INFO nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Took 0.03 seconds to destroy the instance on the hypervisor. [ 763.704426] env[61006]: DEBUG oslo.service.loopingcall [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.704643] env[61006]: DEBUG nova.compute.manager [-] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.704738] env[61006]: DEBUG nova.network.neutron [-] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.731129] env[61006]: DEBUG nova.network.neutron [-] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.929945] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Successfully created port: 8364613d-7b4c-465e-9423-9d702db883a7 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.025020] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.199508] env[61006]: DEBUG nova.network.neutron [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.236838] env[61006]: DEBUG nova.network.neutron [-] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.415071] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41a58a3-0edd-4210-8e4f-7310d73f6219 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.423591] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d26eba-ef38-4447-bd6d-284c5fe6e082 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.457569] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba885a16-fc00-46f7-bf02-52daa086cf7f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.465205] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173ac74a-895c-4bc8-a549-64df1b667dbc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.480564] env[61006]: DEBUG nova.compute.provider_tree [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.576028] env[61006]: DEBUG nova.network.neutron [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.740246] env[61006]: INFO nova.compute.manager [-] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Took 1.04 seconds to deallocate network for instance. [ 764.742689] env[61006]: DEBUG nova.compute.claims [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 764.742840] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.981900] env[61006]: DEBUG nova.scheduler.client.report [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.034545] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.059943] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.060198] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.060349] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.060521] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.060663] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.060849] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.061270] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.061517] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.061701] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.061898] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.062121] env[61006]: DEBUG nova.virt.hardware [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.062974] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e0dcdd-4bf6-46ae-ab37-ef7faa0675c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.073193] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d78c79b-c88f-42b0-90ff-6fa985589a36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.078373] env[61006]: DEBUG oslo_concurrency.lockutils [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] Releasing lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.078604] env[61006]: DEBUG nova.compute.manager [req-7b003586-ea14-4e15-9d88-58f385f68908 req-3a47d055-1479-402b-8ec4-9f377e1f5cff service nova] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Received event network-vif-deleted-f2a13097-5794-4ace-ac92-a25213089df1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.463890] env[61006]: DEBUG nova.compute.manager [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Received event network-changed-8364613d-7b4c-465e-9423-9d702db883a7 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.463890] env[61006]: DEBUG nova.compute.manager [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Refreshing instance network info cache due to event network-changed-8364613d-7b4c-465e-9423-9d702db883a7. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.463890] env[61006]: DEBUG oslo_concurrency.lockutils [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] Acquiring lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.463890] env[61006]: DEBUG oslo_concurrency.lockutils [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] Acquired lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.465784] env[61006]: DEBUG nova.network.neutron [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Refreshing network info cache for port 8364613d-7b4c-465e-9423-9d702db883a7 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.488213] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.488795] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 765.492061] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.343s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.720612] env[61006]: ERROR nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 765.720612] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.720612] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.720612] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.720612] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.720612] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.720612] env[61006]: ERROR nova.compute.manager raise self.value [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.720612] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.720612] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.720612] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.721303] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.721303] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.721303] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 765.721303] env[61006]: ERROR nova.compute.manager [ 765.721303] env[61006]: Traceback (most recent call last): [ 765.721303] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.721303] env[61006]: listener.cb(fileno) [ 765.721303] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.721303] env[61006]: result = function(*args, **kwargs) [ 765.721303] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.721303] env[61006]: return func(*args, **kwargs) [ 765.721303] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.721303] env[61006]: raise e [ 765.721303] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.721303] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 765.721303] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.721303] env[61006]: created_port_ids = self._update_ports_for_instance( [ 765.721303] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.721303] env[61006]: with excutils.save_and_reraise_exception(): [ 765.721303] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.721303] env[61006]: self.force_reraise() [ 765.721303] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.721303] env[61006]: raise self.value [ 765.721303] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.721303] env[61006]: updated_port = self._update_port( [ 765.721303] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.721303] env[61006]: _ensure_no_port_binding_failure(port) [ 765.721303] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.721303] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.722438] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 765.722438] env[61006]: Removing descriptor: 17 [ 765.722438] env[61006]: ERROR nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Traceback (most recent call last): [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] yield resources [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.driver.spawn(context, instance, image_meta, [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.722438] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] vm_ref = self.build_virtual_machine(instance, [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] for vif in network_info: [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self._sync_wrapper(fn, *args, **kwargs) [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.wait() [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self[:] = self._gt.wait() [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self._exit_event.wait() [ 765.722685] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] result = hub.switch() [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self.greenlet.switch() [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] result = function(*args, **kwargs) [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return func(*args, **kwargs) [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise e [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] nwinfo = self.network_api.allocate_for_instance( [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.722953] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] created_port_ids = self._update_ports_for_instance( [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] with excutils.save_and_reraise_exception(): [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.force_reraise() [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise self.value [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] updated_port = self._update_port( [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] _ensure_no_port_binding_failure(port) [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.723275] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise exception.PortBindingFailed(port_id=port['id']) [ 765.724132] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 765.724132] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] [ 765.724132] env[61006]: INFO nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Terminating instance [ 765.728330] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.992595] env[61006]: DEBUG nova.network.neutron [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.995280] env[61006]: DEBUG nova.compute.utils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.996570] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.996738] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 766.087606] env[61006]: DEBUG nova.policy [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '376291781f334b38b700e9921f188105', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'becb5c16e41c422b87349bb527a0b79a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.155116] env[61006]: DEBUG nova.network.neutron [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.403177] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Successfully created port: f52e4e4a-fe88-4787-97b0-262dc67b392d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.504387] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.534693] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance fd415afa-1e5f-42ea-b288-4432f15a5503 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.534823] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4a38192e-4fb8-4767-b1d9-3d8b59a76925 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.534940] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance b8fb9290-755f-40b8-af3d-c17407d0c846 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.535645] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.535645] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0ad1f43f-9e98-43b7-8076-20824ccc18ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.535645] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e4a5a57e-f078-48ec-afe5-204c75139c15 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.535645] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d07742fb-259a-4173-84df-c7c40838ba2c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.535826] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1ebc21ae-d46f-486b-952a-61324aaf353f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.659674] env[61006]: DEBUG oslo_concurrency.lockutils [req-5dbd18fc-9256-484e-8fe2-2f018ab64022 req-a94261f3-d51d-4e21-b370-c840a8c7e738 service nova] Releasing lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.659674] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.659674] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.038963] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance dac63cd3-1788-4e45-b58a-9cd22038dd30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.185089] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.349724] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.493233] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.493474] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.498225] env[61006]: DEBUG nova.compute.manager [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Received event network-vif-deleted-8364613d-7b4c-465e-9423-9d702db883a7 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.498225] env[61006]: DEBUG nova.compute.manager [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Received event network-changed-f52e4e4a-fe88-4787-97b0-262dc67b392d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.498225] env[61006]: DEBUG nova.compute.manager [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Refreshing instance network info cache due to event network-changed-f52e4e4a-fe88-4787-97b0-262dc67b392d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 767.498225] env[61006]: DEBUG oslo_concurrency.lockutils [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] Acquiring lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.498225] env[61006]: DEBUG oslo_concurrency.lockutils [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] Acquired lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.498589] env[61006]: DEBUG nova.network.neutron [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Refreshing network info cache for port f52e4e4a-fe88-4787-97b0-262dc67b392d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.520507] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.545539] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6ee89cb9-ef41-4c85-98d2-5b0190568efe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.560500] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.560655] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.560699] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.561173] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.561173] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.561360] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.561630] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.561630] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.561824] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.562038] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.562214] env[61006]: DEBUG nova.virt.hardware [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.563454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59645c3e-a2dc-4b3d-a546-7555a14eba76 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.574094] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab124d9b-0a55-45ef-b6fe-0d12a1ed88e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.678918] env[61006]: ERROR nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 767.678918] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.678918] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.678918] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.678918] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.678918] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.678918] env[61006]: ERROR nova.compute.manager raise self.value [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.678918] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 767.678918] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.678918] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 767.679550] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.679550] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 767.679550] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 767.679550] env[61006]: ERROR nova.compute.manager [ 767.679550] env[61006]: Traceback (most recent call last): [ 767.679550] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 767.679550] env[61006]: listener.cb(fileno) [ 767.679550] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.679550] env[61006]: result = function(*args, **kwargs) [ 767.679550] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.679550] env[61006]: return func(*args, **kwargs) [ 767.679550] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.679550] env[61006]: raise e [ 767.679550] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.679550] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 767.679550] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.679550] env[61006]: created_port_ids = self._update_ports_for_instance( [ 767.679550] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.679550] env[61006]: with excutils.save_and_reraise_exception(): [ 767.679550] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.679550] env[61006]: self.force_reraise() [ 767.679550] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.679550] env[61006]: raise self.value [ 767.679550] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.679550] env[61006]: updated_port = self._update_port( [ 767.679550] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.679550] env[61006]: _ensure_no_port_binding_failure(port) [ 767.679550] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.679550] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 767.680685] env[61006]: nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 767.680685] env[61006]: Removing descriptor: 20 [ 767.680685] env[61006]: ERROR nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Traceback (most recent call last): [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] yield resources [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.driver.spawn(context, instance, image_meta, [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.680685] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] vm_ref = self.build_virtual_machine(instance, [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] for vif in network_info: [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self._sync_wrapper(fn, *args, **kwargs) [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.wait() [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self[:] = self._gt.wait() [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self._exit_event.wait() [ 767.681179] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] result = hub.switch() [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self.greenlet.switch() [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] result = function(*args, **kwargs) [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return func(*args, **kwargs) [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise e [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] nwinfo = self.network_api.allocate_for_instance( [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.681670] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] created_port_ids = self._update_ports_for_instance( [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] with excutils.save_and_reraise_exception(): [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.force_reraise() [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise self.value [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] updated_port = self._update_port( [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] _ensure_no_port_binding_failure(port) [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.682542] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise exception.PortBindingFailed(port_id=port['id']) [ 767.682860] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 767.682860] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] [ 767.682860] env[61006]: INFO nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Terminating instance [ 767.682860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.852234] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Releasing lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.852652] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.852840] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 767.853209] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f514cfdc-0615-4006-b0f1-37464845e3aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.862809] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b149091d-b966-4d9b-864f-14dc09e00359 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.886019] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d07742fb-259a-4173-84df-c7c40838ba2c could not be found. [ 767.886019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 767.886019] env[61006]: INFO nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 767.886019] env[61006]: DEBUG oslo.service.loopingcall [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.886019] env[61006]: DEBUG nova.compute.manager [-] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.886019] env[61006]: DEBUG nova.network.neutron [-] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.900510] env[61006]: DEBUG nova.network.neutron [-] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.018015] env[61006]: DEBUG nova.network.neutron [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.054043] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance cca6374c-09a4-4145-a116-c49c5a8330c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.110363] env[61006]: DEBUG nova.network.neutron [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.403267] env[61006]: DEBUG nova.network.neutron [-] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.558035] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6e99894e-81b6-4a07-9ec7-caa16272b3ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.614610] env[61006]: DEBUG oslo_concurrency.lockutils [req-e83ee129-273e-4bb2-ba75-c09e80c9b2da req-3dec7be1-5dfa-4e66-8400-cde6330e38d5 service nova] Releasing lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.615041] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.615235] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.906013] env[61006]: INFO nova.compute.manager [-] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Took 1.02 seconds to deallocate network for instance. [ 768.908630] env[61006]: DEBUG nova.compute.claims [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 768.908842] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.061261] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e2a40bd9-fb66-40a2-bcf1-5c74707d59dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.138755] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.220944] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.521429] env[61006]: DEBUG nova.compute.manager [req-13c0bc49-348e-4e4b-8090-a993b028f1bf req-e8c869df-d890-490a-9ef7-3833b116e80b service nova] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Received event network-vif-deleted-f52e4e4a-fe88-4787-97b0-262dc67b392d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.564045] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 88e2bdc9-ab73-4e23-94b5-a45046835144 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.723636] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.724060] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.724255] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.724537] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96d49a92-bc0e-493c-a817-9e02e756b41d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.734124] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53c2b77-16d6-4bcf-a157-9d5eb0775d86 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.754381] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ebc21ae-d46f-486b-952a-61324aaf353f could not be found. [ 769.754597] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.754771] env[61006]: INFO nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 769.755021] env[61006]: DEBUG oslo.service.loopingcall [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.755248] env[61006]: DEBUG nova.compute.manager [-] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.755337] env[61006]: DEBUG nova.network.neutron [-] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.775557] env[61006]: DEBUG nova.network.neutron [-] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.067699] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance ea7d5d49-ac76-4f2e-9456-912cf466fcc2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.277577] env[61006]: DEBUG nova.network.neutron [-] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.570224] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 2a4089c9-4229-40bd-8d0e-706bba94655f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.780983] env[61006]: INFO nova.compute.manager [-] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Took 1.03 seconds to deallocate network for instance. [ 770.783623] env[61006]: DEBUG nova.compute.claims [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 770.783623] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.073776] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42b92d52-d1f0-48ff-94b6-6164b387456c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.576757] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 598c0d72-d679-49a8-b17c-f5f341c205e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.080928] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 29c41817-2189-4622-8f35-86f61eb34bed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.586703] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 72c3e37b-2eac-41d4-8308-0a6466c2dd24 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.090307] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 524f3fd1-1e71-40c0-96c2-0acac5055e01 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.594596] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 28b83ea7-5433-472d-9e47-f73a4f2fb389 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.102105] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0111452e-1b4f-499c-932d-f31364d1a14c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.604196] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance aedaa5d6-e0f2-492c-a14b-3254863e1f06 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.107381] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.610992] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.611383] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 775.611553] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 775.696506] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Refreshing inventories for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 775.711314] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updating ProviderTree inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 775.711542] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 775.723050] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Refreshing aggregate associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, aggregates: None {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 775.740435] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Refreshing trait associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 776.106344] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff630ec0-4df5-4608-b5b5-ba94fd432286 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.113835] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7dd11ed-6373-472e-ad6d-b45249d887fd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.143133] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e801b980-92f1-4712-b34f-f2b65b31796e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.150229] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322f6e93-586a-4f96-a002-851f3c8aaf0f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.163804] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.667065] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.172024] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 777.172363] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.680s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.172530] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.555s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.175353] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 777.175506] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Cleaning up deleted instances {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 777.682395] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] There are 5 instances to clean {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 777.682681] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 47124964-8057-46ba-8c57-1a9f4471402c] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 777.971316] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b365ea3-92e8-4714-95e1-c6dd35218f4c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.979047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa5c205-9b32-448a-aa8d-964487c0200a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.008187] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd42e21-b63a-4e11-9413-582fe10fcb60 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.015401] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a07be3-5a94-485d-a8e1-e3cc3895049f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.027940] env[61006]: DEBUG nova.compute.provider_tree [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.186056] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6642ffcb-cb01-4e38-a27c-bf4e4c938a17] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 778.531170] env[61006]: DEBUG nova.scheduler.client.report [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.689048] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 41826fe1-8f39-479a-b9fd-51399753dfb5] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.035935] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.863s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.037033] env[61006]: ERROR nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Traceback (most recent call last): [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.driver.spawn(context, instance, image_meta, [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] vm_ref = self.build_virtual_machine(instance, [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.037033] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] for vif in network_info: [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self._sync_wrapper(fn, *args, **kwargs) [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.wait() [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self[:] = self._gt.wait() [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self._exit_event.wait() [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] result = hub.switch() [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.037386] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return self.greenlet.switch() [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] result = function(*args, **kwargs) [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] return func(*args, **kwargs) [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise e [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] nwinfo = self.network_api.allocate_for_instance( [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] created_port_ids = self._update_ports_for_instance( [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] with excutils.save_and_reraise_exception(): [ 779.037711] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] self.force_reraise() [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise self.value [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] updated_port = self._update_port( [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] _ensure_no_port_binding_failure(port) [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] raise exception.PortBindingFailed(port_id=port['id']) [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] nova.exception.PortBindingFailed: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. [ 779.038045] env[61006]: ERROR nova.compute.manager [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] [ 779.038345] env[61006]: DEBUG nova.compute.utils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.038973] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.224s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.042677] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Build of instance fd415afa-1e5f-42ea-b288-4432f15a5503 was re-scheduled: Binding failed for port 236ab510-684c-453b-ad81-da8a8d22898d, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.042677] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.042677] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.042677] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.042889] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.192302] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 116d50e2-d8fa-4e26-8301-d1066d627982] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.578943] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.678112] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.696328] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a2360f64-2bed-4c0a-9f99-54b2a34f8d68] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.877766] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e3b3a3-7fca-4ede-9d8b-7f0e8fdaad59 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.885670] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f13c64-477e-4a4c-860b-e06b6e5d61fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.916169] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212a76cb-ba6d-4421-8646-874e38213b4a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.923487] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0770ced9-93d6-474d-9488-fd60dfabdb2d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.938648] env[61006]: DEBUG nova.compute.provider_tree [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.179730] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-fd415afa-1e5f-42ea-b288-4432f15a5503" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.179918] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.180927] env[61006]: DEBUG nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.180927] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.195870] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.198947] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 780.199112] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Cleaning up deleted instances with incomplete migration {{(pid=61006) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 780.442301] env[61006]: DEBUG nova.scheduler.client.report [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.699301] env[61006]: DEBUG nova.network.neutron [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.702202] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 780.946726] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.947371] env[61006]: ERROR nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Traceback (most recent call last): [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.driver.spawn(context, instance, image_meta, [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] vm_ref = self.build_virtual_machine(instance, [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.947371] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] for vif in network_info: [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return self._sync_wrapper(fn, *args, **kwargs) [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.wait() [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self[:] = self._gt.wait() [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return self._exit_event.wait() [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] current.throw(*self._exc) [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.947645] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] result = function(*args, **kwargs) [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] return func(*args, **kwargs) [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise e [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] nwinfo = self.network_api.allocate_for_instance( [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] created_port_ids = self._update_ports_for_instance( [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] with excutils.save_and_reraise_exception(): [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] self.force_reraise() [ 780.947916] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise self.value [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] updated_port = self._update_port( [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] _ensure_no_port_binding_failure(port) [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] raise exception.PortBindingFailed(port_id=port['id']) [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] nova.exception.PortBindingFailed: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. [ 780.948232] env[61006]: ERROR nova.compute.manager [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] [ 780.948232] env[61006]: DEBUG nova.compute.utils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.949351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.626s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.952142] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Build of instance 4a38192e-4fb8-4767-b1d9-3d8b59a76925 was re-scheduled: Binding failed for port b572595f-70ee-4065-87bc-5d4eb2849b52, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.952548] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.953347] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.953508] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.953665] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.201824] env[61006]: INFO nova.compute.manager [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: fd415afa-1e5f-42ea-b288-4432f15a5503] Took 1.02 seconds to deallocate network for instance. [ 781.474023] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.542740] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.737503] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25743f19-c4c1-4bdd-911b-38a145abcf1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.746039] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b2758e-d713-4b75-b505-0fcec7818c9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.777284] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6ddb5d-2015-413e-8550-1975bd25c6c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.785201] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0404bbc-8682-4a3b-8a6f-48d072c68ccf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.799045] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.045690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-4a38192e-4fb8-4767-b1d9-3d8b59a76925" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.045976] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.046185] env[61006]: DEBUG nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.046354] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.061814] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.229971] env[61006]: INFO nova.scheduler.client.report [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocations for instance fd415afa-1e5f-42ea-b288-4432f15a5503 [ 782.302838] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.564638] env[61006]: DEBUG nova.network.neutron [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.737450] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e9b798-a91a-4219-a233-4d414260d32f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "fd415afa-1e5f-42ea-b288-4432f15a5503" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.732s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.807690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.858s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.808362] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Traceback (most recent call last): [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.driver.spawn(context, instance, image_meta, [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] vm_ref = self.build_virtual_machine(instance, [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.808362] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] for vif in network_info: [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return self._sync_wrapper(fn, *args, **kwargs) [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.wait() [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self[:] = self._gt.wait() [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return self._exit_event.wait() [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] current.throw(*self._exc) [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.808704] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] result = function(*args, **kwargs) [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] return func(*args, **kwargs) [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise e [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] nwinfo = self.network_api.allocate_for_instance( [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] created_port_ids = self._update_ports_for_instance( [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] with excutils.save_and_reraise_exception(): [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] self.force_reraise() [ 782.809099] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise self.value [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] updated_port = self._update_port( [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] _ensure_no_port_binding_failure(port) [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] raise exception.PortBindingFailed(port_id=port['id']) [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] nova.exception.PortBindingFailed: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. [ 782.809508] env[61006]: ERROR nova.compute.manager [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] [ 782.809508] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.810382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.067s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.812078] env[61006]: INFO nova.compute.claims [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.814584] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Build of instance b8fb9290-755f-40b8-af3d-c17407d0c846 was re-scheduled: Binding failed for port ce883c90-2640-4e48-be26-5ff7136fda9e, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.815016] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.815649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.815649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.815649] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.068690] env[61006]: INFO nova.compute.manager [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 4a38192e-4fb8-4767-b1d9-3d8b59a76925] Took 1.02 seconds to deallocate network for instance. [ 783.239615] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.343638] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.431970] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.763503] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.935007] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-b8fb9290-755f-40b8-af3d-c17407d0c846" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.935249] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.935428] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.935587] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.950254] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.097594] env[61006]: INFO nova.scheduler.client.report [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Deleted allocations for instance 4a38192e-4fb8-4767-b1d9-3d8b59a76925 [ 784.116563] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cea693-3e95-4d7e-92cb-ed5693a453bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.124759] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c13291-5b8d-421a-8b70-827956c595be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.154117] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76e2ad4-6b0c-4187-9fb1-e70e9179b679 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.160784] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b85315b-e81b-495e-87f5-dd7f18d560f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.173613] env[61006]: DEBUG nova.compute.provider_tree [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.354550] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.354806] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.452540] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.607504] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f90e440f-8c5f-4730-86b8-fc979ecf83d3 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "4a38192e-4fb8-4767-b1d9-3d8b59a76925" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 160.854s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.676506] env[61006]: DEBUG nova.scheduler.client.report [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.954849] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: b8fb9290-755f-40b8-af3d-c17407d0c846] Took 1.02 seconds to deallocate network for instance. [ 785.110230] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.181793] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.181793] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.184153] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.494s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.185545] env[61006]: INFO nova.compute.claims [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.633156] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.689783] env[61006]: DEBUG nova.compute.utils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.692896] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.693082] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 785.742821] env[61006]: DEBUG nova.policy [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3dca34ba528a4fe98ae1bc7b88594395', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0070f033ed4b433cb9109f434cd4f5e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.984137] env[61006]: INFO nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Deleted allocations for instance b8fb9290-755f-40b8-af3d-c17407d0c846 [ 786.045433] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Successfully created port: 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.196283] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.496712] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "b8fb9290-755f-40b8-af3d-c17407d0c846" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.416s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.512694] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab71e5c-fd08-4b1f-9e83-2ab860a639b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.521145] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1746d22c-df8d-4640-93d3-b79ccece9a3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.554721] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae804bdf-00d0-4a70-873e-11cacef3fcb4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.562511] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91b8c22-43c7-41c0-b469-4de8fba77ce6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.575641] env[61006]: DEBUG nova.compute.provider_tree [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.808523] env[61006]: DEBUG nova.compute.manager [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Received event network-changed-9fb779e3-0280-44d5-b9cc-2ef05e6cd56e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.808853] env[61006]: DEBUG nova.compute.manager [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Refreshing instance network info cache due to event network-changed-9fb779e3-0280-44d5-b9cc-2ef05e6cd56e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.809024] env[61006]: DEBUG oslo_concurrency.lockutils [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] Acquiring lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.809154] env[61006]: DEBUG oslo_concurrency.lockutils [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] Acquired lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.809250] env[61006]: DEBUG nova.network.neutron [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Refreshing network info cache for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 786.993121] env[61006]: ERROR nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 786.993121] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.993121] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.993121] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.993121] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.993121] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.993121] env[61006]: ERROR nova.compute.manager raise self.value [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.993121] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 786.993121] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.993121] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 786.993590] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.993590] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 786.993590] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 786.993590] env[61006]: ERROR nova.compute.manager [ 786.993590] env[61006]: Traceback (most recent call last): [ 786.993590] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 786.993590] env[61006]: listener.cb(fileno) [ 786.993590] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.993590] env[61006]: result = function(*args, **kwargs) [ 786.993590] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.993590] env[61006]: return func(*args, **kwargs) [ 786.993590] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.993590] env[61006]: raise e [ 786.993590] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.993590] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 786.993590] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.993590] env[61006]: created_port_ids = self._update_ports_for_instance( [ 786.993590] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.993590] env[61006]: with excutils.save_and_reraise_exception(): [ 786.993590] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.993590] env[61006]: self.force_reraise() [ 786.993590] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.993590] env[61006]: raise self.value [ 786.993590] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.993590] env[61006]: updated_port = self._update_port( [ 786.993590] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.993590] env[61006]: _ensure_no_port_binding_failure(port) [ 786.993590] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.993590] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 786.994406] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 786.994406] env[61006]: Removing descriptor: 20 [ 787.000060] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.079015] env[61006]: DEBUG nova.scheduler.client.report [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.207609] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.239969] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.240263] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.240426] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.240610] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.240752] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.240894] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.241308] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.241308] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.241443] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.241598] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.241764] env[61006]: DEBUG nova.virt.hardware [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.242656] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad305dc-4337-4204-8a3b-cd62cf1269d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.251032] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7a756f-14cb-45cc-8a2a-a6eee5252c3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.264565] env[61006]: ERROR nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Traceback (most recent call last): [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] yield resources [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.driver.spawn(context, instance, image_meta, [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] vm_ref = self.build_virtual_machine(instance, [ 787.264565] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] for vif in network_info: [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return self._sync_wrapper(fn, *args, **kwargs) [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.wait() [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self[:] = self._gt.wait() [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return self._exit_event.wait() [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.264900] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] current.throw(*self._exc) [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] result = function(*args, **kwargs) [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return func(*args, **kwargs) [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise e [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] nwinfo = self.network_api.allocate_for_instance( [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] created_port_ids = self._update_ports_for_instance( [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] with excutils.save_and_reraise_exception(): [ 787.265252] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.force_reraise() [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise self.value [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] updated_port = self._update_port( [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] _ensure_no_port_binding_failure(port) [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise exception.PortBindingFailed(port_id=port['id']) [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 787.265592] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] [ 787.265592] env[61006]: INFO nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Terminating instance [ 787.266803] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquiring lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.326087] env[61006]: DEBUG nova.network.neutron [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.408617] env[61006]: DEBUG nova.network.neutron [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.525023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.583951] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.584528] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.587124] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.966s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.911678] env[61006]: DEBUG oslo_concurrency.lockutils [req-b8bceaa5-8dee-433f-b4a1-c79a5bc16326 req-96828074-c551-4492-adcc-7dbbea9e9990 service nova] Releasing lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.912136] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquired lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.912325] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.091806] env[61006]: DEBUG nova.compute.utils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.097323] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.097445] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.156117] env[61006]: DEBUG nova.policy [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.384824] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35649975-0be9-45bd-96f7-be654ab18590 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.392100] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e755ef2e-3717-4efc-9a68-2f97d6b36732 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.424481] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101baf43-f91f-43b3-bb53-b9f8efbc6f5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.440354] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f8e0f3-93b2-4274-b431-c7fceb1eee9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.455887] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.462216] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.494017] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Successfully created port: 06d2127d-fbef-4b53-acdd-7513e200bff4 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.598921] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.634950] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.839917] env[61006]: DEBUG nova.compute.manager [req-a7c98151-ca05-4400-b01f-3696b51a45e9 req-57622392-4f46-48a8-b1fb-22f843488891 service nova] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Received event network-vif-deleted-9fb779e3-0280-44d5-b9cc-2ef05e6cd56e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.966132] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.137352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Releasing lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.137764] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.137972] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.138322] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b705a655-dff2-4124-9016-48a45e29a910 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.147897] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b75c0a-f9d3-40e0-8055-6ccfb47a8c47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.173540] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dac63cd3-1788-4e45-b58a-9cd22038dd30 could not be found. [ 789.173791] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.173970] env[61006]: INFO nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Took 0.04 seconds to destroy the instance on the hypervisor. [ 789.174255] env[61006]: DEBUG oslo.service.loopingcall [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.174499] env[61006]: DEBUG nova.compute.manager [-] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.174597] env[61006]: DEBUG nova.network.neutron [-] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.193498] env[61006]: DEBUG nova.network.neutron [-] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.471617] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.472296] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Traceback (most recent call last): [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.driver.spawn(context, instance, image_meta, [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] vm_ref = self.build_virtual_machine(instance, [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.472296] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] for vif in network_info: [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return self._sync_wrapper(fn, *args, **kwargs) [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.wait() [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self[:] = self._gt.wait() [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return self._exit_event.wait() [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] current.throw(*self._exc) [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.472621] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] result = function(*args, **kwargs) [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] return func(*args, **kwargs) [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise e [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] nwinfo = self.network_api.allocate_for_instance( [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] created_port_ids = self._update_ports_for_instance( [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] with excutils.save_and_reraise_exception(): [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] self.force_reraise() [ 789.472970] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise self.value [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] updated_port = self._update_port( [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] _ensure_no_port_binding_failure(port) [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] raise exception.PortBindingFailed(port_id=port['id']) [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] nova.exception.PortBindingFailed: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. [ 789.473304] env[61006]: ERROR nova.compute.manager [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] [ 789.473304] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 789.474244] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.087s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.479800] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Build of instance 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a was re-scheduled: Binding failed for port 5b65a13b-f572-4137-80ce-2adecd295af8, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 789.480474] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 789.480927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.481251] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.481683] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.487927] env[61006]: ERROR nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 789.487927] env[61006]: ERROR nova.compute.manager Traceback (most recent call last): [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.487927] env[61006]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.487927] env[61006]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.487927] env[61006]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.487927] env[61006]: ERROR nova.compute.manager self.force_reraise() [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.487927] env[61006]: ERROR nova.compute.manager raise self.value [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.487927] env[61006]: ERROR nova.compute.manager updated_port = self._update_port( [ 789.487927] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.487927] env[61006]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 789.488313] env[61006]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.488313] env[61006]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 789.488313] env[61006]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 789.488313] env[61006]: ERROR nova.compute.manager [ 789.488313] env[61006]: Traceback (most recent call last): [ 789.488313] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 789.488313] env[61006]: listener.cb(fileno) [ 789.488313] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.488313] env[61006]: result = function(*args, **kwargs) [ 789.488313] env[61006]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.488313] env[61006]: return func(*args, **kwargs) [ 789.488313] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.488313] env[61006]: raise e [ 789.488313] env[61006]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.488313] env[61006]: nwinfo = self.network_api.allocate_for_instance( [ 789.488313] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.488313] env[61006]: created_port_ids = self._update_ports_for_instance( [ 789.488313] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.488313] env[61006]: with excutils.save_and_reraise_exception(): [ 789.488313] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.488313] env[61006]: self.force_reraise() [ 789.488313] env[61006]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.488313] env[61006]: raise self.value [ 789.488313] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.488313] env[61006]: updated_port = self._update_port( [ 789.488313] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.488313] env[61006]: _ensure_no_port_binding_failure(port) [ 789.488313] env[61006]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.488313] env[61006]: raise exception.PortBindingFailed(port_id=port['id']) [ 789.489018] env[61006]: nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 789.489018] env[61006]: Removing descriptor: 20 [ 789.613857] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.642275] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.642598] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.642778] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.642960] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.643124] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.643272] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.643536] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.643696] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.643858] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.644026] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.644200] env[61006]: DEBUG nova.virt.hardware [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.645383] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff96374-cc2d-45dd-aa48-18759adb7e42 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.653749] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29db8778-b391-4807-a018-7d951a54fb8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.668794] env[61006]: ERROR nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Traceback (most recent call last): [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] yield resources [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.driver.spawn(context, instance, image_meta, [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] vm_ref = self.build_virtual_machine(instance, [ 789.668794] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] for vif in network_info: [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return self._sync_wrapper(fn, *args, **kwargs) [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.wait() [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self[:] = self._gt.wait() [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return self._exit_event.wait() [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 789.669189] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] current.throw(*self._exc) [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] result = function(*args, **kwargs) [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return func(*args, **kwargs) [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise e [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] nwinfo = self.network_api.allocate_for_instance( [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] created_port_ids = self._update_ports_for_instance( [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] with excutils.save_and_reraise_exception(): [ 789.669566] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.force_reraise() [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise self.value [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] updated_port = self._update_port( [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] _ensure_no_port_binding_failure(port) [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise exception.PortBindingFailed(port_id=port['id']) [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 789.669960] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] [ 789.669960] env[61006]: INFO nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Terminating instance [ 789.672836] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.672836] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.672836] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.696212] env[61006]: DEBUG nova.network.neutron [-] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.998313] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.086510] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.196340] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.200986] env[61006]: INFO nova.compute.manager [-] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Took 1.03 seconds to deallocate network for instance. [ 790.204417] env[61006]: DEBUG nova.compute.claims [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 790.204596] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.283361] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.351755] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd83f6b4-6672-488c-94fd-ccf721f17583 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.362594] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41c6f8d-ca22-454c-aacd-03ee266b20b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.392890] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19244aa8-b25e-447a-bef5-b22188e6e4a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.400430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7f481a-b8e2-4a90-a332-52f052cee2c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.414764] env[61006]: DEBUG nova.compute.provider_tree [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.591300] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.591652] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 790.591852] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.592134] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.608780] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.785371] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.785829] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 790.786069] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 790.786420] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4583e3e-9b7f-45c7-83b1-e05e007dc9cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.797692] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e01d457-3db6-4b74-891f-aa2ef024dd71 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.827052] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ee89cb9-ef41-4c85-98d2-5b0190568efe could not be found. [ 790.827288] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.827496] env[61006]: INFO nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 790.827761] env[61006]: DEBUG oslo.service.loopingcall [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.828012] env[61006]: DEBUG nova.compute.manager [-] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.828087] env[61006]: DEBUG nova.network.neutron [-] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.848527] env[61006]: DEBUG nova.network.neutron [-] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.912843] env[61006]: DEBUG nova.compute.manager [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Received event network-changed-06d2127d-fbef-4b53-acdd-7513e200bff4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.914521] env[61006]: DEBUG nova.compute.manager [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Refreshing instance network info cache due to event network-changed-06d2127d-fbef-4b53-acdd-7513e200bff4. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.914745] env[61006]: DEBUG oslo_concurrency.lockutils [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] Acquiring lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.914905] env[61006]: DEBUG oslo_concurrency.lockutils [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] Acquired lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.915094] env[61006]: DEBUG nova.network.neutron [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Refreshing network info cache for port 06d2127d-fbef-4b53-acdd-7513e200bff4 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.921540] env[61006]: DEBUG nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.112041] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.351688] env[61006]: DEBUG nova.network.neutron [-] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.425648] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.426248] env[61006]: ERROR nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Traceback (most recent call last): [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.driver.spawn(context, instance, image_meta, [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] vm_ref = self.build_virtual_machine(instance, [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.426248] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] for vif in network_info: [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self._sync_wrapper(fn, *args, **kwargs) [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.wait() [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self[:] = self._gt.wait() [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self._exit_event.wait() [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] result = hub.switch() [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 791.426550] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return self.greenlet.switch() [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] result = function(*args, **kwargs) [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] return func(*args, **kwargs) [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise e [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] nwinfo = self.network_api.allocate_for_instance( [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] created_port_ids = self._update_ports_for_instance( [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] with excutils.save_and_reraise_exception(): [ 791.426887] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] self.force_reraise() [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise self.value [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] updated_port = self._update_port( [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] _ensure_no_port_binding_failure(port) [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] raise exception.PortBindingFailed(port_id=port['id']) [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] nova.exception.PortBindingFailed: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. [ 791.427220] env[61006]: ERROR nova.compute.manager [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] [ 791.427502] env[61006]: DEBUG nova.compute.utils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.428281] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.685s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.431948] env[61006]: DEBUG nova.network.neutron [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.433978] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Build of instance 0ad1f43f-9e98-43b7-8076-20824ccc18ed was re-scheduled: Binding failed for port c18dcc73-eae1-4b1f-b584-4f31b9d413f5, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 791.434433] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 791.434649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquiring lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.434788] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Acquired lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.435033] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.505947] env[61006]: DEBUG nova.network.neutron [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.614373] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a] Took 1.02 seconds to deallocate network for instance. [ 791.854567] env[61006]: INFO nova.compute.manager [-] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Took 1.03 seconds to deallocate network for instance. [ 791.856714] env[61006]: DEBUG nova.compute.claims [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Aborting claim: {{(pid=61006) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 791.856888] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.961994] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.009030] env[61006]: DEBUG oslo_concurrency.lockutils [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] Releasing lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.009298] env[61006]: DEBUG nova.compute.manager [req-c20a5140-d8e0-4eb5-b8e2-3fdf27c43612 req-3a849542-aa90-4da6-a418-66d6e9ed2ce6 service nova] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Received event network-vif-deleted-06d2127d-fbef-4b53-acdd-7513e200bff4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.032906] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.205479] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e2da7e-22d2-47b9-a275-a9a9d1890257 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.213618] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76454981-5d0d-4227-952f-0f820870bafd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.243957] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df7e607-64ca-4b71-ac90-1b4b06eaff49 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.251358] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302aef48-90b8-4086-858e-f5c513d4f291 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.264344] env[61006]: DEBUG nova.compute.provider_tree [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.535472] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Releasing lock "refresh_cache-0ad1f43f-9e98-43b7-8076-20824ccc18ed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.535640] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.535820] env[61006]: DEBUG nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.535984] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.645039] env[61006]: INFO nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Deleted allocations for instance 96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a [ 792.665874] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.767703] env[61006]: DEBUG nova.scheduler.client.report [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.156437] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "96fcfe8b-c30b-4a0b-bdbf-9d3eff42eb4a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.044s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.168374] env[61006]: DEBUG nova.network.neutron [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.272495] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.844s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.273138] env[61006]: ERROR nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Traceback (most recent call last): [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.driver.spawn(context, instance, image_meta, [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] vm_ref = self.build_virtual_machine(instance, [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.273138] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] for vif in network_info: [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self._sync_wrapper(fn, *args, **kwargs) [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.wait() [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self[:] = self._gt.wait() [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self._exit_event.wait() [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] result = hub.switch() [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.273456] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return self.greenlet.switch() [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] result = function(*args, **kwargs) [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] return func(*args, **kwargs) [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise e [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] nwinfo = self.network_api.allocate_for_instance( [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] created_port_ids = self._update_ports_for_instance( [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] with excutils.save_and_reraise_exception(): [ 793.273816] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] self.force_reraise() [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise self.value [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] updated_port = self._update_port( [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] _ensure_no_port_binding_failure(port) [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] raise exception.PortBindingFailed(port_id=port['id']) [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] nova.exception.PortBindingFailed: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. [ 793.274210] env[61006]: ERROR nova.compute.manager [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] [ 793.274509] env[61006]: DEBUG nova.compute.utils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.274982] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.366s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.277816] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Build of instance e4a5a57e-f078-48ec-afe5-204c75139c15 was re-scheduled: Binding failed for port f2a13097-5794-4ace-ac92-a25213089df1, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.278250] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.278544] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.278738] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.278955] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.658969] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.672093] env[61006]: INFO nova.compute.manager [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] [instance: 0ad1f43f-9e98-43b7-8076-20824ccc18ed] Took 1.14 seconds to deallocate network for instance. [ 793.803870] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.876926] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.038376] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8a8515-cfa5-45c0-9d59-15aac9cbf7f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.045895] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663f28bc-3d6d-43e7-975d-6c4ffbe99ca7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.076038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25271d47-a3d8-409f-bae1-0c40f5f51ef4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.083028] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0740cd7b-f949-4c6a-b157-e40879f70213 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.096058] env[61006]: DEBUG nova.compute.provider_tree [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.179181] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.379896] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-e4a5a57e-f078-48ec-afe5-204c75139c15" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.379896] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 794.379896] env[61006]: DEBUG nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.380335] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.403979] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.599053] env[61006]: DEBUG nova.scheduler.client.report [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.697436] env[61006]: INFO nova.scheduler.client.report [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Deleted allocations for instance 0ad1f43f-9e98-43b7-8076-20824ccc18ed [ 794.906649] env[61006]: DEBUG nova.network.neutron [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.104248] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.829s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.104677] env[61006]: ERROR nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Traceback (most recent call last): [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.driver.spawn(context, instance, image_meta, [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] vm_ref = self.build_virtual_machine(instance, [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.104677] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] for vif in network_info: [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self._sync_wrapper(fn, *args, **kwargs) [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.wait() [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self[:] = self._gt.wait() [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self._exit_event.wait() [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] result = hub.switch() [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 795.104949] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return self.greenlet.switch() [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] result = function(*args, **kwargs) [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] return func(*args, **kwargs) [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise e [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] nwinfo = self.network_api.allocate_for_instance( [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] created_port_ids = self._update_ports_for_instance( [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] with excutils.save_and_reraise_exception(): [ 795.105243] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] self.force_reraise() [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise self.value [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] updated_port = self._update_port( [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] _ensure_no_port_binding_failure(port) [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] raise exception.PortBindingFailed(port_id=port['id']) [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] nova.exception.PortBindingFailed: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. [ 795.105579] env[61006]: ERROR nova.compute.manager [instance: d07742fb-259a-4173-84df-c7c40838ba2c] [ 795.105871] env[61006]: DEBUG nova.compute.utils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.107088] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.323s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.109531] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Build of instance d07742fb-259a-4173-84df-c7c40838ba2c was re-scheduled: Binding failed for port 8364613d-7b4c-465e-9423-9d702db883a7, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.109943] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.110180] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.110328] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.110487] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.205866] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3ec1e028-99d6-4215-82e4-7a9808cdf4be tempest-ListServersNegativeTestJSON-675675622 tempest-ListServersNegativeTestJSON-675675622-project-member] Lock "0ad1f43f-9e98-43b7-8076-20824ccc18ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 171.050s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.410043] env[61006]: INFO nova.compute.manager [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: e4a5a57e-f078-48ec-afe5-204c75139c15] Took 1.03 seconds to deallocate network for instance. [ 795.632937] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.708532] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.732183] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.906662] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e42d11-35a9-446f-bcbc-77da8ff58873 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.663578] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Releasing lock "refresh_cache-d07742fb-259a-4173-84df-c7c40838ba2c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.663898] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.664022] env[61006]: DEBUG nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.664184] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.672631] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af446413-93ea-43af-adc5-7ba5466fc6a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.703784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.704588] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.706217] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a00e87-e0b9-452f-9f95-63edd88bf66e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.714195] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48530fe6-4b63-49db-b323-c69d097182e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.727589] env[61006]: DEBUG nova.compute.provider_tree [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.188640] env[61006]: INFO nova.scheduler.client.report [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Deleted allocations for instance e4a5a57e-f078-48ec-afe5-204c75139c15 [ 797.210203] env[61006]: DEBUG nova.network.neutron [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.230675] env[61006]: DEBUG nova.scheduler.client.report [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.703382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-86720ce7-30a8-4bf0-95ce-e966dfc384bc tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "e4a5a57e-f078-48ec-afe5-204c75139c15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.259s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.712628] env[61006]: INFO nova.compute.manager [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d07742fb-259a-4173-84df-c7c40838ba2c] Took 1.05 seconds to deallocate network for instance. [ 797.737212] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.630s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.737832] env[61006]: ERROR nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Traceback (most recent call last): [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.driver.spawn(context, instance, image_meta, [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] vm_ref = self.build_virtual_machine(instance, [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.737832] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] for vif in network_info: [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self._sync_wrapper(fn, *args, **kwargs) [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.wait() [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self[:] = self._gt.wait() [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self._exit_event.wait() [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] result = hub.switch() [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 797.738216] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return self.greenlet.switch() [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] result = function(*args, **kwargs) [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] return func(*args, **kwargs) [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise e [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] nwinfo = self.network_api.allocate_for_instance( [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] created_port_ids = self._update_ports_for_instance( [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] with excutils.save_and_reraise_exception(): [ 797.738497] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] self.force_reraise() [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise self.value [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] updated_port = self._update_port( [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] _ensure_no_port_binding_failure(port) [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] raise exception.PortBindingFailed(port_id=port['id']) [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] nova.exception.PortBindingFailed: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. [ 797.738808] env[61006]: ERROR nova.compute.manager [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] [ 797.739068] env[61006]: DEBUG nova.compute.utils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.739689] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.976s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.741238] env[61006]: INFO nova.compute.claims [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.745205] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Build of instance 1ebc21ae-d46f-486b-952a-61324aaf353f was re-scheduled: Binding failed for port f52e4e4a-fe88-4787-97b0-262dc67b392d, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.745605] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.745823] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquiring lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.745963] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Acquired lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.746132] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.205968] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.281659] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.413633] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.736488] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.748014] env[61006]: INFO nova.scheduler.client.report [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Deleted allocations for instance d07742fb-259a-4173-84df-c7c40838ba2c [ 798.915887] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Releasing lock "refresh_cache-1ebc21ae-d46f-486b-952a-61324aaf353f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.915887] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 798.915887] env[61006]: DEBUG nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.916063] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.933172] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.050783] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1172d695-d44c-4c00-8dfa-eee752c88a00 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.058303] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d39b0a9-ba85-474e-86fa-cea2fa49692f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.088875] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb17167-835b-4978-9099-92ab5b93a373 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.096232] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969bd22d-cf9d-48f8-9bb7-735ba47e1627 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.110152] env[61006]: DEBUG nova.compute.provider_tree [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.255610] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e78ac763-2b3d-4615-86c7-d39715db38ed tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d07742fb-259a-4173-84df-c7c40838ba2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 174.219s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.436440] env[61006]: DEBUG nova.network.neutron [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.614738] env[61006]: DEBUG nova.scheduler.client.report [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.759133] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.943281] env[61006]: INFO nova.compute.manager [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] [instance: 1ebc21ae-d46f-486b-952a-61324aaf353f] Took 1.02 seconds to deallocate network for instance. [ 800.120928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.121534] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.125473] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.492s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.129020] env[61006]: INFO nova.compute.claims [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.284047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.634654] env[61006]: DEBUG nova.compute.utils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.637086] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.637373] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.681186] env[61006]: DEBUG nova.policy [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.979489] env[61006]: INFO nova.scheduler.client.report [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Deleted allocations for instance 1ebc21ae-d46f-486b-952a-61324aaf353f [ 801.059621] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Successfully created port: 598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.140837] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.395152] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.395388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.480941] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baf7ea7-df47-414b-9150-43c6aca98371 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.489741] env[61006]: DEBUG oslo_concurrency.lockutils [None req-811aab83-8113-4aa4-b989-4ba43331f7b1 tempest-ListServerFiltersTestJSON-309410936 tempest-ListServerFiltersTestJSON-309410936-project-member] Lock "1ebc21ae-d46f-486b-952a-61324aaf353f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 176.258s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.492089] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be4195e-3fda-419a-a303-464352b72d79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.537087] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc385b1-142a-4eaa-a9a1-96cf53a94e0f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.543938] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f290da66-0143-4eaf-8483-489d65ce343b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.558116] env[61006]: DEBUG nova.compute.provider_tree [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.000063] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.060921] env[61006]: DEBUG nova.scheduler.client.report [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.153754] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.177713] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.177959] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.178128] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.178308] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.178444] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.178583] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.178774] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.178921] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.179126] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.179307] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.179472] env[61006]: DEBUG nova.virt.hardware [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.180704] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80c3f75-52ba-4f88-91ab-d34a3dd4bdef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.188266] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485014f9-bbfc-4fcf-98a2-90dad93b9f1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.485903] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.486146] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.521388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.567494] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.567494] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.569770] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.045s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.572466] env[61006]: INFO nova.compute.claims [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.812292] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Successfully updated port: 598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.818530] env[61006]: DEBUG nova.compute.manager [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-plugged-598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.818748] env[61006]: DEBUG oslo_concurrency.lockutils [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.818999] env[61006]: DEBUG oslo_concurrency.lockutils [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.819179] env[61006]: DEBUG oslo_concurrency.lockutils [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.819347] env[61006]: DEBUG nova.compute.manager [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] No waiting events found dispatching network-vif-plugged-598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.819506] env[61006]: WARNING nova.compute.manager [req-24434d9c-b540-491a-b76e-b3aa11714ff0 req-d366bdb5-042e-4bc6-8e07-5df6e43c5c7c service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received unexpected event network-vif-plugged-598c1248-f434-4f05-ac91-c7f1322879fa for instance with vm_state building and task_state spawning. [ 803.074023] env[61006]: DEBUG nova.compute.utils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.074023] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.077328] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.162803] env[61006]: DEBUG nova.policy [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6729893a0de46bd96c12a87b9f913be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7461286a17024522bce95d6e78ab2c88', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.321237] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.321420] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.321580] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.581040] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Successfully created port: 9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.583379] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.878126] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.909504] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16d5e37-3c4c-49fb-8383-63b54fb22428 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.924188] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1535c1c1-5244-4b62-b1c1-2cd4f5ce5c4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.956821] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff08999-f0df-4925-8f8f-9f03c8482be2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.966435] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda10251-8ecf-47e2-94f0-80b732eff5ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.982849] env[61006]: DEBUG nova.compute.provider_tree [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.103760] env[61006]: DEBUG nova.network.neutron [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.482506] env[61006]: DEBUG nova.scheduler.client.report [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.597461] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.606143] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.606500] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Instance network_info: |[{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.606977] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:67:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '598c1248-f434-4f05-ac91-c7f1322879fa', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.614985] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Creating folder: Project (85b03e26e0034e30b74761724d0a39e3). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.615275] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27b4903b-2739-4884-bacd-40e605399342 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.625197] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.625881] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.625881] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.625881] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.625881] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.626013] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.626224] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.626385] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.626516] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.626672] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.626873] env[61006]: DEBUG nova.virt.hardware [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.627960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9aea92-ac81-4ef9-9195-69fb420a25c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.631630] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Created folder: Project (85b03e26e0034e30b74761724d0a39e3) in parent group-v285275. [ 804.631799] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Creating folder: Instances. Parent ref: group-v285297. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.632477] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-271fcd3e-e7a4-4087-b4d9-a024fcad1299 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.637196] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dd8814-83fc-4f25-a0ba-162c313d82cf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.651575] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Created folder: Instances in parent group-v285297. [ 804.651838] env[61006]: DEBUG oslo.service.loopingcall [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.651969] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.652198] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e365095-25a7-45ec-a396-d294579c1c13 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.671010] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.671010] env[61006]: value = "task-1336915" [ 804.671010] env[61006]: _type = "Task" [ 804.671010] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.679119] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336915, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.841309] env[61006]: DEBUG nova.compute.manager [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-changed-598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.841501] env[61006]: DEBUG nova.compute.manager [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing instance network info cache due to event network-changed-598c1248-f434-4f05-ac91-c7f1322879fa. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.841704] env[61006]: DEBUG oslo_concurrency.lockutils [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.841844] env[61006]: DEBUG oslo_concurrency.lockutils [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.842009] env[61006]: DEBUG nova.network.neutron [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing network info cache for port 598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.987719] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.988261] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.991192] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.786s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.185198] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336915, 'name': CreateVM_Task, 'duration_secs': 0.325544} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.185504] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.206701] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.206899] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.207873] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.207873] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29ae7711-9e72-4d05-a71e-e9186312e7ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.213119] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 805.213119] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52775bb1-e41d-463d-d6f6-dc943a061dc4" [ 805.213119] env[61006]: _type = "Task" [ 805.213119] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.221352] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52775bb1-e41d-463d-d6f6-dc943a061dc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.283853] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Successfully updated port: 9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.496708] env[61006]: DEBUG nova.compute.utils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.500750] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.500918] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.564517] env[61006]: DEBUG nova.policy [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82aea0ad3a245af99868fc53a0b04f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd26132a67fc1412c938a36094a34a433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.725819] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52775bb1-e41d-463d-d6f6-dc943a061dc4, 'name': SearchDatastore_Task, 'duration_secs': 0.011744} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.726256] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.726570] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.726809] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.726954] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.727177] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.727442] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68224403-a71c-4273-a688-96b7336252d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.737590] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.738555] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.738555] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-917beb8f-bf43-42b0-a818-48f7e0467c85 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.744806] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 805.744806] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ccd03-e415-f207-922d-6bb2b9c14d84" [ 805.744806] env[61006]: _type = "Task" [ 805.744806] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.752752] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ccd03-e415-f207-922d-6bb2b9c14d84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.780723] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ff5f8c-2aa2-4a08-b0ae-335ce60a98e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.787307] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.787547] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquired lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.787638] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.789448] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3ac2ba-363a-4809-bb8c-5dae56e4dcd6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.794533] env[61006]: DEBUG nova.network.neutron [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updated VIF entry in instance network info cache for port 598c1248-f434-4f05-ac91-c7f1322879fa. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.794699] env[61006]: DEBUG nova.network.neutron [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.822777] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb579b3-0d47-4093-a239-a30d48cdecee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.830678] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecc72a5-6f0e-4c72-b88e-a3a8f01953e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.846943] env[61006]: DEBUG nova.compute.provider_tree [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.004343] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.104077] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Successfully created port: fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.260813] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ccd03-e415-f207-922d-6bb2b9c14d84, 'name': SearchDatastore_Task, 'duration_secs': 0.013026} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.261641] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-591bb303-1346-4348-bd77-f343ed06e82f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.267096] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 806.267096] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98de8-35a4-b3ff-2223-d28e796ef8b4" [ 806.267096] env[61006]: _type = "Task" [ 806.267096] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.274868] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98de8-35a4-b3ff-2223-d28e796ef8b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.298796] env[61006]: DEBUG oslo_concurrency.lockutils [req-d8b88b81-5e20-49b4-8c66-ce665c930162 req-41530293-df59-4707-bc3e-095aeaec19ae service nova] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.336054] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.347838] env[61006]: DEBUG nova.scheduler.client.report [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.521121] env[61006]: DEBUG nova.network.neutron [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updating instance_info_cache with network_info: [{"id": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "address": "fa:16:3e:94:26:4f", "network": {"id": "b5daed40-610a-4542-9f27-25017f74bee8", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1202407684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7461286a17024522bce95d6e78ab2c88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e1bc75a-8b", "ovs_interfaceid": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.777796] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98de8-35a4-b3ff-2223-d28e796ef8b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010042} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.778021] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.778241] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] cca6374c-09a4-4145-a116-c49c5a8330c1/cca6374c-09a4-4145-a116-c49c5a8330c1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.778711] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cd6c26f-88b2-43ec-ab27-a1c1b580b09f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.786782] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 806.786782] env[61006]: value = "task-1336916" [ 806.786782] env[61006]: _type = "Task" [ 806.786782] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.794420] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336916, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.853266] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.862s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.853946] env[61006]: ERROR nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Traceback (most recent call last): [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.driver.spawn(context, instance, image_meta, [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] vm_ref = self.build_virtual_machine(instance, [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.853946] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] for vif in network_info: [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return self._sync_wrapper(fn, *args, **kwargs) [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.wait() [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self[:] = self._gt.wait() [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return self._exit_event.wait() [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] current.throw(*self._exc) [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.854306] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] result = function(*args, **kwargs) [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] return func(*args, **kwargs) [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise e [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] nwinfo = self.network_api.allocate_for_instance( [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] created_port_ids = self._update_ports_for_instance( [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] with excutils.save_and_reraise_exception(): [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] self.force_reraise() [ 806.854591] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise self.value [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] updated_port = self._update_port( [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] _ensure_no_port_binding_failure(port) [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] raise exception.PortBindingFailed(port_id=port['id']) [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] nova.exception.PortBindingFailed: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. [ 806.854891] env[61006]: ERROR nova.compute.manager [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] [ 806.854891] env[61006]: DEBUG nova.compute.utils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 806.855978] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.999s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.858967] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Build of instance dac63cd3-1788-4e45-b58a-9cd22038dd30 was re-scheduled: Binding failed for port 9fb779e3-0280-44d5-b9cc-2ef05e6cd56e, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 806.859494] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 806.859669] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquiring lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.859813] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Acquired lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.859965] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.876433] env[61006]: DEBUG nova.compute.manager [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Received event network-vif-plugged-9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.876631] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Acquiring lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.876825] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.876992] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.877171] env[61006]: DEBUG nova.compute.manager [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] No waiting events found dispatching network-vif-plugged-9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.877446] env[61006]: WARNING nova.compute.manager [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Received unexpected event network-vif-plugged-9e1bc75a-8bdf-4cfb-ba15-044d90816855 for instance with vm_state building and task_state spawning. [ 806.877514] env[61006]: DEBUG nova.compute.manager [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Received event network-changed-9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.877623] env[61006]: DEBUG nova.compute.manager [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Refreshing instance network info cache due to event network-changed-9e1bc75a-8bdf-4cfb-ba15-044d90816855. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.877776] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Acquiring lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.014324] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.024492] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Releasing lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.024948] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Instance network_info: |[{"id": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "address": "fa:16:3e:94:26:4f", "network": {"id": "b5daed40-610a-4542-9f27-25017f74bee8", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1202407684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7461286a17024522bce95d6e78ab2c88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e1bc75a-8b", "ovs_interfaceid": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.025296] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Acquired lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.025477] env[61006]: DEBUG nova.network.neutron [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Refreshing network info cache for port 9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.026938] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:26:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd7b5f1ef-d4b9-4ec3-b047-17e4cb349d25', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e1bc75a-8bdf-4cfb-ba15-044d90816855', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.036317] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Creating folder: Project (7461286a17024522bce95d6e78ab2c88). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.040369] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ffed80a-aed3-49d9-8e9a-64a7d1503abf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.053638] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.054010] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.054239] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.054586] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.055158] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.055158] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.055404] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.057404] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.057404] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.057404] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.057404] env[61006]: DEBUG nova.virt.hardware [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.057802] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836c4452-0d7b-464f-a382-3634bfacc9de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.062747] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Created folder: Project (7461286a17024522bce95d6e78ab2c88) in parent group-v285275. [ 807.063036] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Creating folder: Instances. Parent ref: group-v285300. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.063896] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f3eb1fa-15b2-49e7-8c47-34bc67d06af3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.071013] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086d7b5f-7898-4619-bfee-0fc2de229ded {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.080075] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Created folder: Instances in parent group-v285300. [ 807.080458] env[61006]: DEBUG oslo.service.loopingcall [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.081229] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.081774] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e98125e-2f97-4987-8af7-907e5394aa56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.113019] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.113019] env[61006]: value = "task-1336919" [ 807.113019] env[61006]: _type = "Task" [ 807.113019] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.121715] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336919, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.297023] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336916, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.359612] env[61006]: DEBUG nova.network.neutron [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updated VIF entry in instance network info cache for port 9e1bc75a-8bdf-4cfb-ba15-044d90816855. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 807.359998] env[61006]: DEBUG nova.network.neutron [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updating instance_info_cache with network_info: [{"id": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "address": "fa:16:3e:94:26:4f", "network": {"id": "b5daed40-610a-4542-9f27-25017f74bee8", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1202407684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7461286a17024522bce95d6e78ab2c88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e1bc75a-8b", "ovs_interfaceid": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.390169] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.539941] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.624838] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336919, 'name': CreateVM_Task, 'duration_secs': 0.463975} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.627799] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.630582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.630748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.631058] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.632190] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07450257-0925-4322-bdc5-fa713de0e4d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.635947] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 807.635947] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52799c99-6256-4c33-66a2-9bd45dfa31b7" [ 807.635947] env[61006]: _type = "Task" [ 807.635947] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.646087] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52799c99-6256-4c33-66a2-9bd45dfa31b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.694198] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c5a927-a4e9-4fb5-bc6e-4ba54cea50d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.701097] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be27d22-f0a8-4bb2-80da-a0a95c48c754 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.729505] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7b7328-8f11-4029-aa46-92608b7ff295 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.736412] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823e7298-68f9-4da6-84a8-0f4f282ae1de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.748971] env[61006]: DEBUG nova.compute.provider_tree [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.797723] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336916, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513269} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.797924] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] cca6374c-09a4-4145-a116-c49c5a8330c1/cca6374c-09a4-4145-a116-c49c5a8330c1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.798147] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.798382] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-783340a8-0fae-445c-aaad-61cc3e4c0817 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.805417] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 807.805417] env[61006]: value = "task-1336920" [ 807.805417] env[61006]: _type = "Task" [ 807.805417] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.809304] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Successfully updated port: fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.813341] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.868210] env[61006]: DEBUG oslo_concurrency.lockutils [req-7edc7c4b-54a9-4c10-aa29-e69083998236 req-a36bf7e9-ae17-4e80-aa8d-0ddf8292a3a3 service nova] Releasing lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.043267] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Releasing lock "refresh_cache-dac63cd3-1788-4e45-b58a-9cd22038dd30" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.044677] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.044937] env[61006]: DEBUG nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.045134] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.066367] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.146083] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52799c99-6256-4c33-66a2-9bd45dfa31b7, 'name': SearchDatastore_Task, 'duration_secs': 0.010185} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.146389] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.146616] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.146838] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.146980] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.147180] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.147430] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47e7a30d-7b43-438c-8405-f0606b0ad543 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.155125] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.155292] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.155974] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cbfb64d-1f8a-44cd-b48c-c6ed5a7b32df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.160383] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 808.160383] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c85060-3685-9812-c519-00467a87aa87" [ 808.160383] env[61006]: _type = "Task" [ 808.160383] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.169441] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c85060-3685-9812-c519-00467a87aa87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.252196] env[61006]: DEBUG nova.scheduler.client.report [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.315448] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.315718] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.315718] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.316801] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067465} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.317227] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.317979] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28386e4-b194-4e78-8ead-fd941d92480d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.340382] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] cca6374c-09a4-4145-a116-c49c5a8330c1/cca6374c-09a4-4145-a116-c49c5a8330c1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.340753] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3745e17-011f-4f50-908a-9d76fe385415 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.359522] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 808.359522] env[61006]: value = "task-1336921" [ 808.359522] env[61006]: _type = "Task" [ 808.359522] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.367148] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.568655] env[61006]: DEBUG nova.network.neutron [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.670943] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c85060-3685-9812-c519-00467a87aa87, 'name': SearchDatastore_Task, 'duration_secs': 0.012064} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.673155] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a57d9f9-d6c2-490e-a681-d76d3ef3c137 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.676668] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 808.676668] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52298fff-8d80-569e-a4c5-e797358983e7" [ 808.676668] env[61006]: _type = "Task" [ 808.676668] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.686523] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52298fff-8d80-569e-a4c5-e797358983e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.757417] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.901s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.757830] env[61006]: ERROR nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Traceback (most recent call last): [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.driver.spawn(context, instance, image_meta, [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] vm_ref = self.build_virtual_machine(instance, [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] vif_infos = vmwarevif.get_vif_info(self._session, [ 808.757830] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] for vif in network_info: [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return self._sync_wrapper(fn, *args, **kwargs) [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.wait() [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self[:] = self._gt.wait() [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return self._exit_event.wait() [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] current.throw(*self._exc) [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.758249] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] result = function(*args, **kwargs) [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] return func(*args, **kwargs) [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise e [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] nwinfo = self.network_api.allocate_for_instance( [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] created_port_ids = self._update_ports_for_instance( [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] with excutils.save_and_reraise_exception(): [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] self.force_reraise() [ 808.758545] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise self.value [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] updated_port = self._update_port( [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] _ensure_no_port_binding_failure(port) [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] raise exception.PortBindingFailed(port_id=port['id']) [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] nova.exception.PortBindingFailed: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. [ 808.758838] env[61006]: ERROR nova.compute.manager [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] [ 808.758838] env[61006]: DEBUG nova.compute.utils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 808.759787] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.581s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.761335] env[61006]: INFO nova.compute.claims [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.764782] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Build of instance 6ee89cb9-ef41-4c85-98d2-5b0190568efe was re-scheduled: Binding failed for port 06d2127d-fbef-4b53-acdd-7513e200bff4, please check neutron logs for more information. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 808.764782] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Unplugging VIFs for instance {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 808.764782] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.764963] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.764999] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.845217] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.870583] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336921, 'name': ReconfigVM_Task, 'duration_secs': 0.313048} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.870583] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfigured VM instance instance-0000003a to attach disk [datastore2] cca6374c-09a4-4145-a116-c49c5a8330c1/cca6374c-09a4-4145-a116-c49c5a8330c1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.870761] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d0f3454-7957-4855-84aa-4c18eb5db418 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.877336] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 808.877336] env[61006]: value = "task-1336922" [ 808.877336] env[61006]: _type = "Task" [ 808.877336] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.884863] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336922, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.902683] env[61006]: DEBUG nova.compute.manager [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Received event network-vif-plugged-fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.902847] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Acquiring lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.903049] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.903220] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.903402] env[61006]: DEBUG nova.compute.manager [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] No waiting events found dispatching network-vif-plugged-fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.903680] env[61006]: WARNING nova.compute.manager [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Received unexpected event network-vif-plugged-fcaabd7b-dda4-4f87-abae-d9d5335eee51 for instance with vm_state building and task_state spawning. [ 808.903815] env[61006]: DEBUG nova.compute.manager [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Received event network-changed-fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.903988] env[61006]: DEBUG nova.compute.manager [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Refreshing instance network info cache due to event network-changed-fcaabd7b-dda4-4f87-abae-d9d5335eee51. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.904587] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Acquiring lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.001595] env[61006]: DEBUG nova.network.neutron [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Updating instance_info_cache with network_info: [{"id": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "address": "fa:16:3e:d7:39:35", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcaabd7b-dd", "ovs_interfaceid": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.071185] env[61006]: INFO nova.compute.manager [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] [instance: dac63cd3-1788-4e45-b58a-9cd22038dd30] Took 1.03 seconds to deallocate network for instance. [ 809.186708] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52298fff-8d80-569e-a4c5-e797358983e7, 'name': SearchDatastore_Task, 'duration_secs': 0.009865} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.186967] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.187237] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6e99894e-81b6-4a07-9ec7-caa16272b3ba/6e99894e-81b6-4a07-9ec7-caa16272b3ba.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.187479] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1091e91d-7848-49b0-964e-92baa9a64a01 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.194042] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 809.194042] env[61006]: value = "task-1336923" [ 809.194042] env[61006]: _type = "Task" [ 809.194042] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.200724] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336923, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.283222] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.385034] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.390276] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336922, 'name': Rename_Task, 'duration_secs': 0.137269} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.390866] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.391092] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43927a07-9deb-42e4-b16c-a46435b75331 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.398257] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 809.398257] env[61006]: value = "task-1336924" [ 809.398257] env[61006]: _type = "Task" [ 809.398257] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.407648] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336924, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.504655] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.505031] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Instance network_info: |[{"id": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "address": "fa:16:3e:d7:39:35", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcaabd7b-dd", "ovs_interfaceid": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.505428] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Acquired lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.505610] env[61006]: DEBUG nova.network.neutron [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Refreshing network info cache for port fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.506927] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:39:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1931669-8959-4e86-a603-e206bcf2b47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcaabd7b-dda4-4f87-abae-d9d5335eee51', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.514684] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating folder: Project (d26132a67fc1412c938a36094a34a433). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.515928] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-531af98a-cc5b-457d-9db4-6a94016085e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.527336] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created folder: Project (d26132a67fc1412c938a36094a34a433) in parent group-v285275. [ 809.527477] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating folder: Instances. Parent ref: group-v285303. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.527813] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2cfbb766-4163-434d-aae7-07e9b4f35c0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.536606] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created folder: Instances in parent group-v285303. [ 809.536853] env[61006]: DEBUG oslo.service.loopingcall [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.537059] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.537276] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb70d8d7-bd68-4e2d-907e-39e65bf94122 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.557625] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.557625] env[61006]: value = "task-1336927" [ 809.557625] env[61006]: _type = "Task" [ 809.557625] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.565952] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336927, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.703527] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336923, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485242} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.703527] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6e99894e-81b6-4a07-9ec7-caa16272b3ba/6e99894e-81b6-4a07-9ec7-caa16272b3ba.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.703708] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.703971] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e8b7423-05ce-4514-a385-1206d84efc4e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.711421] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 809.711421] env[61006]: value = "task-1336928" [ 809.711421] env[61006]: _type = "Task" [ 809.711421] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.722521] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336928, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.891704] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-6ee89cb9-ef41-4c85-98d2-5b0190568efe" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.891930] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61006) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 809.892150] env[61006]: DEBUG nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.892379] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.908978] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336924, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.909825] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.038603] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e736386f-823a-435f-bdbe-5f8e98d9c799 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.046081] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c3858b-36b4-479b-8dca-e99d1f0a29c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.084285] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f6662f-c40f-465c-b71f-dedfa0a22fa1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.097414] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440a1e0d-5fdc-4519-af91-c9220094639a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.101168] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336927, 'name': CreateVM_Task, 'duration_secs': 0.337467} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.102100] env[61006]: INFO nova.scheduler.client.report [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Deleted allocations for instance dac63cd3-1788-4e45-b58a-9cd22038dd30 [ 810.107991] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.110563] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.110719] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.111022] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.112222] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-376ffd12-2861-4bd7-a56d-abb680de38fb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.123317] env[61006]: DEBUG nova.compute.provider_tree [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.128609] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 810.128609] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52729cf3-39b4-f29e-2eb2-9b775a6e0602" [ 810.128609] env[61006]: _type = "Task" [ 810.128609] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.141040] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52729cf3-39b4-f29e-2eb2-9b775a6e0602, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.220680] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336928, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060736} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.220959] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.224431] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26ca19b-96e7-4844-9d11-1fddf8907b64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.244097] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 6e99894e-81b6-4a07-9ec7-caa16272b3ba/6e99894e-81b6-4a07-9ec7-caa16272b3ba.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.244415] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a0cf1e6-cda4-4e3b-aad8-558a112f0bdc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.259468] env[61006]: DEBUG nova.network.neutron [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Updated VIF entry in instance network info cache for port fcaabd7b-dda4-4f87-abae-d9d5335eee51. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 810.259997] env[61006]: DEBUG nova.network.neutron [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Updating instance_info_cache with network_info: [{"id": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "address": "fa:16:3e:d7:39:35", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcaabd7b-dd", "ovs_interfaceid": "fcaabd7b-dda4-4f87-abae-d9d5335eee51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.266997] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 810.266997] env[61006]: value = "task-1336929" [ 810.266997] env[61006]: _type = "Task" [ 810.266997] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.277351] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336929, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.411092] env[61006]: DEBUG oslo_vmware.api [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1336924, 'name': PowerOnVM_Task, 'duration_secs': 0.588671} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.411463] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.411622] env[61006]: INFO nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Took 8.26 seconds to spawn the instance on the hypervisor. [ 810.411849] env[61006]: DEBUG nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.412669] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115f9760-856f-4632-b89e-7accf97909a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.415414] env[61006]: DEBUG nova.network.neutron [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.615048] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f2443fd2-39bd-439c-a3aa-2e731a2b8726 tempest-ServersNegativeTestJSON-697398050 tempest-ServersNegativeTestJSON-697398050-project-member] Lock "dac63cd3-1788-4e45-b58a-9cd22038dd30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 177.689s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.626694] env[61006]: DEBUG nova.scheduler.client.report [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.640825] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52729cf3-39b4-f29e-2eb2-9b775a6e0602, 'name': SearchDatastore_Task, 'duration_secs': 0.009435} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.641068] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.641318] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.641854] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.642011] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.642243] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.643090] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-514c2e1a-ed06-4724-9250-201c7245e984 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.654039] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.654217] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.654897] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fa7fa54-229f-4395-953f-68d17d413c6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.659994] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 810.659994] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522adaa1-b7d7-9086-69b2-0506e47e49b7" [ 810.659994] env[61006]: _type = "Task" [ 810.659994] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.667680] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522adaa1-b7d7-9086-69b2-0506e47e49b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.762947] env[61006]: DEBUG oslo_concurrency.lockutils [req-33673ee5-cefc-4f7e-8177-af2abf871838 req-51f13879-e6a7-418c-86c4-91a824ced5e4 service nova] Releasing lock "refresh_cache-e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.777191] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336929, 'name': ReconfigVM_Task, 'duration_secs': 0.265019} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.777470] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 6e99894e-81b6-4a07-9ec7-caa16272b3ba/6e99894e-81b6-4a07-9ec7-caa16272b3ba.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.778094] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fbbd463-c645-4aa7-bdea-d8f0eee5debb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.784718] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 810.784718] env[61006]: value = "task-1336930" [ 810.784718] env[61006]: _type = "Task" [ 810.784718] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.792500] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336930, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.918384] env[61006]: INFO nova.compute.manager [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 6ee89cb9-ef41-4c85-98d2-5b0190568efe] Took 1.03 seconds to deallocate network for instance. [ 810.932842] env[61006]: INFO nova.compute.manager [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Took 27.19 seconds to build instance. [ 811.120676] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.134021] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.134021] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.134535] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.431s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.139333] env[61006]: INFO nova.compute.claims [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.179022] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522adaa1-b7d7-9086-69b2-0506e47e49b7, 'name': SearchDatastore_Task, 'duration_secs': 0.008698} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.179022] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b352e24a-d89a-465b-aa49-2fc050d2253f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.183373] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 811.183373] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52772c10-5f37-c63f-3e8d-6a4094286fc4" [ 811.183373] env[61006]: _type = "Task" [ 811.183373] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.193605] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52772c10-5f37-c63f-3e8d-6a4094286fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.296342] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336930, 'name': Rename_Task, 'duration_secs': 0.138667} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.296342] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.296342] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6afa844b-8716-4341-8680-f6f4eae5b09e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.303020] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 811.303020] env[61006]: value = "task-1336931" [ 811.303020] env[61006]: _type = "Task" [ 811.303020] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.316114] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.436375] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9e9dbf8e-3cb3-4c09-b654-75a44269e8e4 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.984s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.453152] env[61006]: DEBUG nova.compute.manager [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-changed-598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.453993] env[61006]: DEBUG nova.compute.manager [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing instance network info cache due to event network-changed-598c1248-f434-4f05-ac91-c7f1322879fa. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.454430] env[61006]: DEBUG oslo_concurrency.lockutils [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.454742] env[61006]: DEBUG oslo_concurrency.lockutils [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.455100] env[61006]: DEBUG nova.network.neutron [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing network info cache for port 598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.644289] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.645982] env[61006]: DEBUG nova.compute.utils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.652018] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.652018] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 811.695243] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52772c10-5f37-c63f-3e8d-6a4094286fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.008541} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.695470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.695718] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] e2a40bd9-fb66-40a2-bcf1-5c74707d59dd/e2a40bd9-fb66-40a2-bcf1-5c74707d59dd.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.696048] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71163151-87af-419b-a918-3ed217e6b8f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.703114] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 811.703114] env[61006]: value = "task-1336932" [ 811.703114] env[61006]: _type = "Task" [ 811.703114] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.711433] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336932, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.713026] env[61006]: DEBUG nova.policy [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6c5fe37a73b40b48e7cbb4f4eaa3d74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '064b30c918ac499e9838488a08256d29', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.812946] env[61006]: DEBUG oslo_vmware.api [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1336931, 'name': PowerOnVM_Task, 'duration_secs': 0.452301} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.813233] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.813434] env[61006]: INFO nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Took 7.22 seconds to spawn the instance on the hypervisor. [ 811.813608] env[61006]: DEBUG nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.814729] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb1b85d-ddad-4f5c-abc5-b1898b60e89c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.939857] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.957473] env[61006]: INFO nova.scheduler.client.report [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance 6ee89cb9-ef41-4c85-98d2-5b0190568efe [ 812.150892] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.216098] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336932, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502306} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.218841] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] e2a40bd9-fb66-40a2-bcf1-5c74707d59dd/e2a40bd9-fb66-40a2-bcf1-5c74707d59dd.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.219079] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.219530] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-572b1e57-b205-4534-bf79-255888963765 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.226255] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 812.226255] env[61006]: value = "task-1336933" [ 812.226255] env[61006]: _type = "Task" [ 812.226255] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.238194] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336933, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.334980] env[61006]: INFO nova.compute.manager [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Took 26.72 seconds to build instance. [ 812.461625] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144e6633-6148-4819-bffc-fa636437d4e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.466212] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.466806] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d11f288a-ad18-408d-aa7c-ed394f704ed1 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "6ee89cb9-ef41-4c85-98d2-5b0190568efe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 178.367s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.472915] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49441d2d-1a90-4418-a61d-1a162d8c0f38 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.516750] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c809f4a4-18ae-4aae-8c13-66e5bc358dd7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.527259] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61baa487-f35b-4f45-a071-f4b2be1882c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.545192] env[61006]: DEBUG nova.compute.provider_tree [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.645861] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Successfully created port: e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.736816] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336933, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069739} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.737180] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.737949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cc28bc-b363-4e5f-8fb8-3c6ea411b369 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.761080] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] e2a40bd9-fb66-40a2-bcf1-5c74707d59dd/e2a40bd9-fb66-40a2-bcf1-5c74707d59dd.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.761472] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab442dba-5e6e-4bb1-ad7c-450b2eb33f6e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.781684] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 812.781684] env[61006]: value = "task-1336934" [ 812.781684] env[61006]: _type = "Task" [ 812.781684] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.792435] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336934, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.799704] env[61006]: DEBUG nova.network.neutron [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updated VIF entry in instance network info cache for port 598c1248-f434-4f05-ac91-c7f1322879fa. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.800085] env[61006]: DEBUG nova.network.neutron [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.837398] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ec9089fe-d21e-40e2-98b7-562eaac7690a tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 173.688s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.970509] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.047436] env[61006]: DEBUG nova.scheduler.client.report [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.171876] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.280372] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.280820] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.281065] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.281300] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.281453] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.281600] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.281842] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.282344] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.282539] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.282733] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.282867] env[61006]: DEBUG nova.virt.hardware [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.283840] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58eb157-28e5-4332-8ac9-cc88141225ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.300564] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0451588-623c-42e8-9013-e3c4ac4edf67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.305089] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336934, 'name': ReconfigVM_Task, 'duration_secs': 0.281806} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.305560] env[61006]: DEBUG oslo_concurrency.lockutils [req-f97b5f52-ef97-4fbf-a036-2197a9a9c95b req-c671dba9-23d8-4e0a-9619-733341f74e84 service nova] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.305951] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Reconfigured VM instance instance-0000003c to attach disk [datastore2] e2a40bd9-fb66-40a2-bcf1-5c74707d59dd/e2a40bd9-fb66-40a2-bcf1-5c74707d59dd.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.307149] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42500dbe-4d42-4218-813f-9a4933e0e0d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.322888] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 813.322888] env[61006]: value = "task-1336935" [ 813.322888] env[61006]: _type = "Task" [ 813.322888] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.331714] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336935, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.340244] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.510379] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.553868] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.554532] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.557359] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.821s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.560091] env[61006]: INFO nova.compute.claims [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.834137] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336935, 'name': Rename_Task, 'duration_secs': 0.158065} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.834355] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.834712] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-417fee9c-240e-4d17-975a-07c79dafcaec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.841588] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 813.841588] env[61006]: value = "task-1336936" [ 813.841588] env[61006]: _type = "Task" [ 813.841588] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.855114] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.865208] env[61006]: DEBUG nova.compute.manager [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Received event network-changed-9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.865412] env[61006]: DEBUG nova.compute.manager [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Refreshing instance network info cache due to event network-changed-9e1bc75a-8bdf-4cfb-ba15-044d90816855. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.865669] env[61006]: DEBUG oslo_concurrency.lockutils [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] Acquiring lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.865848] env[61006]: DEBUG oslo_concurrency.lockutils [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] Acquired lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.866018] env[61006]: DEBUG nova.network.neutron [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Refreshing network info cache for port 9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.877426] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.069903] env[61006]: DEBUG nova.compute.utils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.071468] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.071635] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.112559] env[61006]: DEBUG nova.policy [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15985b95557e4a53ad3cb8cd3b89b127', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4640bc05c60f46bab2b3a6e3b2d98cf6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.134505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.134505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.353365] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336936, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.577068] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.613700] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Successfully created port: 8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.650046] env[61006]: DEBUG nova.compute.manager [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received event network-vif-plugged-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.650332] env[61006]: DEBUG oslo_concurrency.lockutils [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] Acquiring lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.650574] env[61006]: DEBUG oslo_concurrency.lockutils [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.650773] env[61006]: DEBUG oslo_concurrency.lockutils [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.651861] env[61006]: DEBUG nova.compute.manager [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] No waiting events found dispatching network-vif-plugged-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.652102] env[61006]: WARNING nova.compute.manager [req-22679bf9-c2b1-415c-bfee-c09cd0d1ca07 req-c3b700cb-f930-4e4c-b6e5-1907f8920331 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received unexpected event network-vif-plugged-e8f20725-aaab-4cc2-87d5-89688205c617 for instance with vm_state building and task_state spawning. [ 814.855305] env[61006]: DEBUG oslo_vmware.api [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336936, 'name': PowerOnVM_Task, 'duration_secs': 0.601953} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.855539] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.855734] env[61006]: INFO nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Took 7.84 seconds to spawn the instance on the hypervisor. [ 814.855904] env[61006]: DEBUG nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.859881] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f88c5b-8812-4ef0-8b99-c70a9ed78e61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.910699] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5de9764-5c37-4a7f-8341-d37b9bc8b1ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.919852] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd46a17e-600e-480a-b908-71385dab921e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.956682] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Successfully updated port: e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.956682] env[61006]: DEBUG nova.network.neutron [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updated VIF entry in instance network info cache for port 9e1bc75a-8bdf-4cfb-ba15-044d90816855. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 814.956747] env[61006]: DEBUG nova.network.neutron [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updating instance_info_cache with network_info: [{"id": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "address": "fa:16:3e:94:26:4f", "network": {"id": "b5daed40-610a-4542-9f27-25017f74bee8", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1202407684-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7461286a17024522bce95d6e78ab2c88", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d7b5f1ef-d4b9-4ec3-b047-17e4cb349d25", "external-id": "nsx-vlan-transportzone-743", "segmentation_id": 743, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e1bc75a-8b", "ovs_interfaceid": "9e1bc75a-8bdf-4cfb-ba15-044d90816855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.956851] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698a63f2-25e6-406f-9bd6-b7cc7bfe299b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.965312] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ce7e73-7621-4b73-bddf-21d21475bb1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.982327] env[61006]: DEBUG nova.compute.provider_tree [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.381549] env[61006]: INFO nova.compute.manager [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Took 27.88 seconds to build instance. [ 815.460238] env[61006]: DEBUG oslo_concurrency.lockutils [req-fe3f5865-a743-489f-b4a4-51ebcfac0f3f req-44f54734-b1eb-490b-bd46-43e880ba1f41 service nova] Releasing lock "refresh_cache-6e99894e-81b6-4a07-9ec7-caa16272b3ba" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.460867] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.460997] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.461160] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.485222] env[61006]: DEBUG nova.scheduler.client.report [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.588560] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.609559] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.609788] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.609941] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.610131] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.610277] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.610494] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.610716] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.610880] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.611071] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.611273] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.611460] env[61006]: DEBUG nova.virt.hardware [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.612336] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d23c990-5ed5-44fb-88d1-8fb4c250ec2d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.620191] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c1bcb7-8d40-405f-84b5-d9ab41b10842 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.883724] env[61006]: DEBUG oslo_concurrency.lockutils [None req-87bd5def-1174-4cd2-905a-3ba50a099166 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.428s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.998331] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.998802] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.001321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.717s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.002947] env[61006]: INFO nova.compute.claims [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.092426] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.301437] env[61006]: DEBUG nova.network.neutron [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.386885] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.401647] env[61006]: DEBUG nova.compute.manager [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Received event network-vif-plugged-8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.401857] env[61006]: DEBUG oslo_concurrency.lockutils [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] Acquiring lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.402075] env[61006]: DEBUG oslo_concurrency.lockutils [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.402271] env[61006]: DEBUG oslo_concurrency.lockutils [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.402469] env[61006]: DEBUG nova.compute.manager [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] No waiting events found dispatching network-vif-plugged-8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.402637] env[61006]: WARNING nova.compute.manager [req-07ccce85-2468-45c2-9530-e0ec59614bd3 req-8c6d3b94-82bb-427e-9abf-86708ff0f8f8 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Received unexpected event network-vif-plugged-8dbbd99a-28ee-4db1-bf70-63d8827606c2 for instance with vm_state building and task_state spawning. [ 816.502569] env[61006]: DEBUG nova.compute.manager [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.503426] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f2692a-6a6e-4500-b312-eab6ff5e9e82 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.507228] env[61006]: DEBUG nova.compute.utils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.510240] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.510407] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 816.555267] env[61006]: DEBUG nova.policy [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5be98dcb34124b6ba2c12e3fc365ec21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '472b5a9c304740fe951a3c501e6db1e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.673518] env[61006]: DEBUG nova.compute.manager [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.673962] env[61006]: DEBUG nova.compute.manager [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing instance network info cache due to event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.673962] env[61006]: DEBUG oslo_concurrency.lockutils [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] Acquiring lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.803847] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.804199] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Instance network_info: |[{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.804500] env[61006]: DEBUG oslo_concurrency.lockutils [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] Acquired lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.804711] env[61006]: DEBUG nova.network.neutron [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.805884] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:b1:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8f20725-aaab-4cc2-87d5-89688205c617', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.814668] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Creating folder: Project (064b30c918ac499e9838488a08256d29). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.815123] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1ab2428-a696-4a78-a22c-80ee6f5a6ff3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.827408] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Created folder: Project (064b30c918ac499e9838488a08256d29) in parent group-v285275. [ 816.827505] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Creating folder: Instances. Parent ref: group-v285306. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.828086] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce766178-acbf-462f-b7b2-8fd2f3810b01 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.837430] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Created folder: Instances in parent group-v285306. [ 816.837678] env[61006]: DEBUG oslo.service.loopingcall [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.837868] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.838088] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b34cc189-1a1f-4806-880c-2f6e212b1527 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.853591] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Successfully created port: a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.860275] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.860275] env[61006]: value = "task-1336939" [ 816.860275] env[61006]: _type = "Task" [ 816.860275] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.867730] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336939, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.910959] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.925860] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Successfully updated port: 8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.014621] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.021914] env[61006]: INFO nova.compute.manager [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] instance snapshotting [ 817.024838] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cd0baa-4a8d-4502-95e9-a61e84350fcc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.045863] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de416531-17fd-4aa3-96be-55b502098417 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.329155] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d34313-2311-4b0d-bf62-91d4fa6f33c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.338676] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a191aae-d474-4fb9-b278-2816781c71c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.371875] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b60adc-26b6-4e7c-9821-a2d63e910809 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.379504] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336939, 'name': CreateVM_Task, 'duration_secs': 0.326569} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.381432] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.382170] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.382380] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.382692] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.383866] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a21b08a-9187-4188-ba0d-25160a3a3f99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.387465] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39ac0887-ea5a-4123-8fb1-2af5a8f17ad7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.398807] env[61006]: DEBUG nova.compute.provider_tree [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.401226] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 817.401226] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5255666b-3828-12ec-c9ec-ec3d3b6a04b5" [ 817.401226] env[61006]: _type = "Task" [ 817.401226] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.412088] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5255666b-3828-12ec-c9ec-ec3d3b6a04b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.430035] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.430035] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquired lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.430035] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.562993] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 817.563255] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8a1efaa3-49c1-49ed-8c1d-e946eeafe074 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.566952] env[61006]: DEBUG nova.network.neutron [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updated VIF entry in instance network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.567297] env[61006]: DEBUG nova.network.neutron [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.575958] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 817.575958] env[61006]: value = "task-1336940" [ 817.575958] env[61006]: _type = "Task" [ 817.575958] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.584407] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336940, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.903192] env[61006]: DEBUG nova.scheduler.client.report [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.918656] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5255666b-3828-12ec-c9ec-ec3d3b6a04b5, 'name': SearchDatastore_Task, 'duration_secs': 0.033356} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.919596] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.919868] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.920148] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.920301] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.920506] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.921062] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57d0468c-cf93-4d19-a5e3-cd1e77009b28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.929630] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.929831] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.930584] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db2eac44-6de9-48e8-9d83-5c39c037bbd8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.938049] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 817.938049] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f1db89-7d2d-3ee4-b32e-c38423c62157" [ 817.938049] env[61006]: _type = "Task" [ 817.938049] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.946115] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f1db89-7d2d-3ee4-b32e-c38423c62157, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.969197] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.028993] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.050249] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.050469] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.050625] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.050804] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.050942] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.051095] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.051368] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.051536] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.051699] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.051855] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.052032] env[61006]: DEBUG nova.virt.hardware [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.053417] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a57dc98-31ad-4304-8f54-7acf79fb8eec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.061200] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4010e26b-4a63-4832-9f68-efa4d19fb17b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.074487] env[61006]: DEBUG oslo_concurrency.lockutils [req-3cc714cc-f274-4286-8ad6-dfae3403c279 req-8a2e4640-797b-4dc3-8cae-cde6c658ea97 service nova] Releasing lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.083598] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336940, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.141487] env[61006]: DEBUG nova.network.neutron [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Updating instance_info_cache with network_info: [{"id": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "address": "fa:16:3e:4b:e2:8d", "network": {"id": "bc145378-7502-4c46-bc95-e9881a5f73de", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1925687963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4640bc05c60f46bab2b3a6e3b2d98cf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dbbd99a-28", "ovs_interfaceid": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.413131] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.413654] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.416326] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.895s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.420047] env[61006]: INFO nova.compute.claims [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.449582] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f1db89-7d2d-3ee4-b32e-c38423c62157, 'name': SearchDatastore_Task, 'duration_secs': 0.009513} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.450204] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d32fad4a-e310-4f2c-bb60-123dff8505c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.456169] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 818.456169] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5286711b-da8c-002d-d586-2721cd9b3bf3" [ 818.456169] env[61006]: _type = "Task" [ 818.456169] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.465213] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5286711b-da8c-002d-d586-2721cd9b3bf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.499008] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Successfully updated port: a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.585702] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336940, 'name': CreateSnapshot_Task, 'duration_secs': 0.530216} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.586033] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 818.586765] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49085f3c-be1d-432a-b94e-a6df4f68801b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.644196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Releasing lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.644521] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Instance network_info: |[{"id": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "address": "fa:16:3e:4b:e2:8d", "network": {"id": "bc145378-7502-4c46-bc95-e9881a5f73de", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1925687963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4640bc05c60f46bab2b3a6e3b2d98cf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dbbd99a-28", "ovs_interfaceid": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.644941] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:e2:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8dbbd99a-28ee-4db1-bf70-63d8827606c2', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.654432] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Creating folder: Project (4640bc05c60f46bab2b3a6e3b2d98cf6). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.654432] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f0cd043-2168-4d6c-a2e8-9694a3a6c417 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.665456] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Created folder: Project (4640bc05c60f46bab2b3a6e3b2d98cf6) in parent group-v285275. [ 818.665632] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Creating folder: Instances. Parent ref: group-v285310. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.665854] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c79b2bf-8b83-48a8-a30a-58653a321c97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.674430] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Created folder: Instances in parent group-v285310. [ 818.675321] env[61006]: DEBUG oslo.service.loopingcall [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.675321] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.675321] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de03734d-95d4-47bc-9db8-517eb7ae83de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.695574] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.695574] env[61006]: value = "task-1336943" [ 818.695574] env[61006]: _type = "Task" [ 818.695574] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.701080] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Received event network-changed-8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.701321] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Refreshing instance network info cache due to event network-changed-8dbbd99a-28ee-4db1-bf70-63d8827606c2. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.701549] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Acquiring lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.701696] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Acquired lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.701839] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Refreshing network info cache for port 8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.705882] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336943, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.921599] env[61006]: DEBUG nova.compute.utils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.925059] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.925059] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.962431] env[61006]: DEBUG nova.policy [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e135b91f4f4541e4b985be02f969e00c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24ba46b1b91f46479841ea9db9dce16f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.969590] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5286711b-da8c-002d-d586-2721cd9b3bf3, 'name': SearchDatastore_Task, 'duration_secs': 0.016503} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.969828] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.970103] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 88e2bdc9-ab73-4e23-94b5-a45046835144/88e2bdc9-ab73-4e23-94b5-a45046835144.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.970386] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88ebfbef-ff83-40a8-853b-2e904dc54c76 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.977224] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 818.977224] env[61006]: value = "task-1336944" [ 818.977224] env[61006]: _type = "Task" [ 818.977224] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.984422] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.001373] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.002024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquired lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.002024] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.105323] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 819.105667] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0ece2adb-0fa8-4475-8a63-001f2fadcb58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.114399] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 819.114399] env[61006]: value = "task-1336945" [ 819.114399] env[61006]: _type = "Task" [ 819.114399] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.122880] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336945, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.209722] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336943, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.428152] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.491773] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336944, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.616280] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Successfully created port: 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.633437] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336945, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.699349] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.713800] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336943, 'name': CreateVM_Task, 'duration_secs': 0.708546} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.713921] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.714495] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.714651] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.714953] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.717495] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5db95b99-c1ee-4c5b-ba93-82300d4fabf8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.724565] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 819.724565] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52949e7b-fb06-ec3f-a4d6-fec7b49e6118" [ 819.724565] env[61006]: _type = "Task" [ 819.724565] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.733095] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52949e7b-fb06-ec3f-a4d6-fec7b49e6118, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.761100] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40995c76-1723-421c-8ea8-fa1d5bf66a8a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.768712] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd28fac-e7f8-4247-8979-32d4cba88053 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.801152] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b201ad3-8d44-4001-bd0d-523867c13aa1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.811149] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2700ad59-c020-4981-bf18-9ec4692ef12a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.826702] env[61006]: DEBUG nova.compute.provider_tree [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.915727] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Updated VIF entry in instance network info cache for port 8dbbd99a-28ee-4db1-bf70-63d8827606c2. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.916166] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Updating instance_info_cache with network_info: [{"id": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "address": "fa:16:3e:4b:e2:8d", "network": {"id": "bc145378-7502-4c46-bc95-e9881a5f73de", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1925687963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4640bc05c60f46bab2b3a6e3b2d98cf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dbbd99a-28", "ovs_interfaceid": "8dbbd99a-28ee-4db1-bf70-63d8827606c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.990871] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688572} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.991201] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 88e2bdc9-ab73-4e23-94b5-a45046835144/88e2bdc9-ab73-4e23-94b5-a45046835144.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.991366] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.991620] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0da8d914-00c8-4412-b0ac-6346d1cabfb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.997720] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 819.997720] env[61006]: value = "task-1336946" [ 819.997720] env[61006]: _type = "Task" [ 819.997720] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.006929] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.040391] env[61006]: DEBUG nova.network.neutron [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Updating instance_info_cache with network_info: [{"id": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "address": "fa:16:3e:41:7b:17", "network": {"id": "1cfff00d-d8a6-4be2-860f-33e5ac7db90b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1823734830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "472b5a9c304740fe951a3c501e6db1e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa94adb0b-5e", "ovs_interfaceid": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.132798] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336945, 'name': CloneVM_Task} progress is 95%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.235009] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52949e7b-fb06-ec3f-a4d6-fec7b49e6118, 'name': SearchDatastore_Task, 'duration_secs': 0.012027} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.235314] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.235543] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.235850] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.236014] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.236202] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.236451] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9264078-76cf-4f16-b98a-c7847e31f957 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.244023] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.244199] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.244897] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce9a1681-1564-4623-b3a0-ddaa1e5348b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.249383] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 820.249383] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270f084-507b-b110-5d6c-7630e3df38d9" [ 820.249383] env[61006]: _type = "Task" [ 820.249383] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.256511] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270f084-507b-b110-5d6c-7630e3df38d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.330155] env[61006]: DEBUG nova.scheduler.client.report [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.418650] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Releasing lock "refresh_cache-ea7d5d49-ac76-4f2e-9456-912cf466fcc2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.418912] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Received event network-vif-plugged-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.419120] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Acquiring lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.419369] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.419538] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.419700] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] No waiting events found dispatching network-vif-plugged-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.419860] env[61006]: WARNING nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Received unexpected event network-vif-plugged-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda for instance with vm_state building and task_state spawning. [ 820.420029] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Received event network-changed-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.420183] env[61006]: DEBUG nova.compute.manager [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Refreshing instance network info cache due to event network-changed-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.420437] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Acquiring lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.441014] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.463520] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.463780] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.463932] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.464122] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.464266] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.464410] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.464612] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.464766] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.464925] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.465092] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.465262] env[61006]: DEBUG nova.virt.hardware [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.466183] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf069363-07dd-41d5-abd6-5bff74d4e3f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.474471] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4128869-7b62-429d-882f-f1d51a6bcfb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.505423] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068966} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.505676] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.506431] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339d88aa-c986-44c8-937d-55665dc2085d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.527446] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 88e2bdc9-ab73-4e23-94b5-a45046835144/88e2bdc9-ab73-4e23-94b5-a45046835144.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.527708] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2df64939-8007-4fd8-ba08-d8612105afb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.543357] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Releasing lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.543809] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Instance network_info: |[{"id": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "address": "fa:16:3e:41:7b:17", "network": {"id": "1cfff00d-d8a6-4be2-860f-33e5ac7db90b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1823734830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "472b5a9c304740fe951a3c501e6db1e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa94adb0b-5e", "ovs_interfaceid": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.544191] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Acquired lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.544360] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Refreshing network info cache for port a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.545549] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:7b:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604c9724-b4ef-4393-a76e-eb4a2b510796', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a94adb0b-5ee2-4af8-bfcf-f268d54f1eda', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.552724] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Creating folder: Project (472b5a9c304740fe951a3c501e6db1e8). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.557624] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-baeec0bd-aa02-4d1a-b756-0ebb7ebbdd20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.559490] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 820.559490] env[61006]: value = "task-1336947" [ 820.559490] env[61006]: _type = "Task" [ 820.559490] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.569443] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.570765] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Created folder: Project (472b5a9c304740fe951a3c501e6db1e8) in parent group-v285275. [ 820.570942] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Creating folder: Instances. Parent ref: group-v285314. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.571194] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b5d1d22-d03f-4236-8c51-639d4c769113 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.581061] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Created folder: Instances in parent group-v285314. [ 820.581246] env[61006]: DEBUG oslo.service.loopingcall [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.581456] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.581811] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8849b8a0-fe7d-4d93-81ca-1eb16a63f351 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.601482] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.601482] env[61006]: value = "task-1336950" [ 820.601482] env[61006]: _type = "Task" [ 820.601482] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.608856] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336950, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.628180] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336945, 'name': CloneVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.762925] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270f084-507b-b110-5d6c-7630e3df38d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009013} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.763894] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b8b0e8-a760-4373-a713-3e26ead3b54f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.769576] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 820.769576] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527acfc5-4272-814f-5c1b-481d20b86b8c" [ 820.769576] env[61006]: _type = "Task" [ 820.769576] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.777665] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527acfc5-4272-814f-5c1b-481d20b86b8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.786212] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Updated VIF entry in instance network info cache for port a94adb0b-5ee2-4af8-bfcf-f268d54f1eda. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.786509] env[61006]: DEBUG nova.network.neutron [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Updating instance_info_cache with network_info: [{"id": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "address": "fa:16:3e:41:7b:17", "network": {"id": "1cfff00d-d8a6-4be2-860f-33e5ac7db90b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1823734830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "472b5a9c304740fe951a3c501e6db1e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa94adb0b-5e", "ovs_interfaceid": "a94adb0b-5ee2-4af8-bfcf-f268d54f1eda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.835797] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.835797] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.838646] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.194s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.840318] env[61006]: INFO nova.compute.claims [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.070030] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.110917] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336950, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.127973] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336945, 'name': CloneVM_Task, 'duration_secs': 1.649524} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.128269] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Created linked-clone VM from snapshot [ 821.129048] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5225a151-3b81-4bb1-b310-25dfd3e26d66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.136785] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Uploading image 06d8ddf9-1620-4813-97cd-7b51789db884 {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 821.160695] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 821.160695] env[61006]: value = "vm-285313" [ 821.160695] env[61006]: _type = "VirtualMachine" [ 821.160695] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 821.161058] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1f33f2e1-ff57-4ad2-b519-bb014f6eafa4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.168463] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lease: (returnval){ [ 821.168463] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52812661-cfb4-629c-15f2-30494a33948e" [ 821.168463] env[61006]: _type = "HttpNfcLease" [ 821.168463] env[61006]: } obtained for exporting VM: (result){ [ 821.168463] env[61006]: value = "vm-285313" [ 821.168463] env[61006]: _type = "VirtualMachine" [ 821.168463] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 821.168708] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the lease: (returnval){ [ 821.168708] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52812661-cfb4-629c-15f2-30494a33948e" [ 821.168708] env[61006]: _type = "HttpNfcLease" [ 821.168708] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 821.175292] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 821.175292] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52812661-cfb4-629c-15f2-30494a33948e" [ 821.175292] env[61006]: _type = "HttpNfcLease" [ 821.175292] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 821.282102] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527acfc5-4272-814f-5c1b-481d20b86b8c, 'name': SearchDatastore_Task, 'duration_secs': 0.021984} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.282333] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.282988] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] ea7d5d49-ac76-4f2e-9456-912cf466fcc2/ea7d5d49-ac76-4f2e-9456-912cf466fcc2.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.283190] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58d4f80f-4a33-439b-a169-af71b9206945 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.289269] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd4385f6-0c97-4aa7-a32e-4b15bd7452b7 req-823c3202-ee4c-4ea3-8f06-8bd9697c133d service nova] Releasing lock "refresh_cache-2a4089c9-4229-40bd-8d0e-706bba94655f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.292538] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 821.292538] env[61006]: value = "task-1336952" [ 821.292538] env[61006]: _type = "Task" [ 821.292538] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.301485] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336952, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.320580] env[61006]: DEBUG nova.compute.manager [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Received event network-vif-plugged-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.320843] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] Acquiring lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.320995] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.321176] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.321555] env[61006]: DEBUG nova.compute.manager [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] No waiting events found dispatching network-vif-plugged-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.321819] env[61006]: WARNING nova.compute.manager [req-8d49eedf-99f5-4f51-b1c4-4456df4bb99e req-c7b7d3e7-5c82-4da3-93a7-b98e5edc4299 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Received unexpected event network-vif-plugged-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a for instance with vm_state building and task_state spawning. [ 821.345287] env[61006]: DEBUG nova.compute.utils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.351666] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.351666] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.397170] env[61006]: DEBUG nova.policy [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42027a2dce454ca98e295dadca850eee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2b5a4bf20a40ee9b8688935de75dc1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.573806] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336947, 'name': ReconfigVM_Task, 'duration_secs': 0.838305} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.574246] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 88e2bdc9-ab73-4e23-94b5-a45046835144/88e2bdc9-ab73-4e23-94b5-a45046835144.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.575124] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5342775-6806-411a-b321-e8ef1eb67cda {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.583641] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 821.583641] env[61006]: value = "task-1336953" [ 821.583641] env[61006]: _type = "Task" [ 821.583641] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.593055] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336953, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.616208] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336950, 'name': CreateVM_Task, 'duration_secs': 0.766875} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.616208] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.616208] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.616208] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.616208] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.616412] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee516f43-ea8d-451e-baa3-e24b2c32d328 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.621177] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 821.621177] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38915-4710-709e-efeb-26d0489eb789" [ 821.621177] env[61006]: _type = "Task" [ 821.621177] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.632049] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38915-4710-709e-efeb-26d0489eb789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.678038] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 821.678038] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52812661-cfb4-629c-15f2-30494a33948e" [ 821.678038] env[61006]: _type = "HttpNfcLease" [ 821.678038] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 821.678038] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 821.678038] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52812661-cfb4-629c-15f2-30494a33948e" [ 821.678038] env[61006]: _type = "HttpNfcLease" [ 821.678038] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 821.679046] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a62c18-fda0-48dd-9e7f-d887c1a5d343 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.692821] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 821.693249] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 821.780502] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Successfully created port: b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.802785] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336952, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.852590] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.862470] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-12bc36c1-ef26-4c79-ae2b-b6fe7e72ea94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.920056] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Successfully updated port: 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.955657] env[61006]: DEBUG nova.compute.manager [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Received event network-changed-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.955932] env[61006]: DEBUG nova.compute.manager [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Refreshing instance network info cache due to event network-changed-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 821.955967] env[61006]: DEBUG oslo_concurrency.lockutils [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] Acquiring lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.956101] env[61006]: DEBUG oslo_concurrency.lockutils [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] Acquired lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.956262] env[61006]: DEBUG nova.network.neutron [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Refreshing network info cache for port 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.102662] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336953, 'name': Rename_Task, 'duration_secs': 0.330105} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.102957] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 822.103195] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0754bbb-84a1-4d39-ac28-8858c40ccf89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.112210] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 822.112210] env[61006]: value = "task-1336954" [ 822.112210] env[61006]: _type = "Task" [ 822.112210] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.120427] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.139133] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38915-4710-709e-efeb-26d0489eb789, 'name': SearchDatastore_Task, 'duration_secs': 0.067245} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.139691] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.139989] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.140433] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.142901] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.142901] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.142901] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79738e01-e8bf-47c8-9cf7-b7e75852a48e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.155345] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.155555] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.156359] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cdf42c1-5977-453c-85cc-5f0a78ec20c4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.161759] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 822.161759] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f0f566-bb95-8756-fb76-1ee883ebe147" [ 822.161759] env[61006]: _type = "Task" [ 822.161759] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.173040] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f0f566-bb95-8756-fb76-1ee883ebe147, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.227777] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e780856b-e66d-4f4e-ba75-b653bd132a89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.236047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d86bea-7074-4742-957e-df41215d1572 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.269298] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723a73c3-1b56-4673-a621-a1b1b23d7eef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.277034] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2450f53a-b027-4d04-8aa8-6ef633f88adf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.292438] env[61006]: DEBUG nova.compute.provider_tree [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.303622] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336952, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601285} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.305213] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] ea7d5d49-ac76-4f2e-9456-912cf466fcc2/ea7d5d49-ac76-4f2e-9456-912cf466fcc2.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.305403] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.306350] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75ea23c4-fe9f-422b-bebc-d8f9b8aefec9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.314843] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 822.314843] env[61006]: value = "task-1336955" [ 822.314843] env[61006]: _type = "Task" [ 822.314843] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.323568] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336955, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.424018] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.492814] env[61006]: DEBUG nova.network.neutron [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.604532] env[61006]: DEBUG nova.network.neutron [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.624691] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336954, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.675209] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f0f566-bb95-8756-fb76-1ee883ebe147, 'name': SearchDatastore_Task, 'duration_secs': 0.010949} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.676139] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f951951-c525-4605-9239-82b9e62380c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.683021] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 822.683021] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a00cfb-f372-565a-8fd9-07cb352948ad" [ 822.683021] env[61006]: _type = "Task" [ 822.683021] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.692753] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a00cfb-f372-565a-8fd9-07cb352948ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.799848] env[61006]: DEBUG nova.scheduler.client.report [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.829170] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336955, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073895} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.830035] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.830560] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd4a1ec-222e-4a2f-a5f0-d9dc2da1e41c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.855097] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] ea7d5d49-ac76-4f2e-9456-912cf466fcc2/ea7d5d49-ac76-4f2e-9456-912cf466fcc2.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.855539] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e33a55db-1d57-4871-bb3a-c8471d6df33a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.871375] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.879042] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 822.879042] env[61006]: value = "task-1336956" [ 822.879042] env[61006]: _type = "Task" [ 822.879042] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.887659] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336956, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.895441] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.895676] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.895850] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.896124] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.896342] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.896542] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.896803] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.897031] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.897230] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.897439] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.897699] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.898539] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d062ad-c545-491c-bbb9-a09b713c2500 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.906619] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0760006-4249-4786-9ec6-4b9d2af77c25 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.107801] env[61006]: DEBUG oslo_concurrency.lockutils [req-8c734fd7-6092-4be1-aa87-70a933e28c3c req-81f54a7c-c455-4971-adfd-aeed25cd8b4e service nova] Releasing lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.108196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquired lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.108408] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.127598] env[61006]: DEBUG oslo_vmware.api [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336954, 'name': PowerOnVM_Task, 'duration_secs': 0.54582} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.128112] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.128457] env[61006]: INFO nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Took 9.96 seconds to spawn the instance on the hypervisor. [ 823.128845] env[61006]: DEBUG nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.130000] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7487fa7a-aa83-4f7c-8c48-c55ebf6587ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.194038] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a00cfb-f372-565a-8fd9-07cb352948ad, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.194258] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.194524] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 2a4089c9-4229-40bd-8d0e-706bba94655f/2a4089c9-4229-40bd-8d0e-706bba94655f.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.194785] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c941a8c8-7de3-483f-8df9-d3accefcfeb6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.203295] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 823.203295] env[61006]: value = "task-1336957" [ 823.203295] env[61006]: _type = "Task" [ 823.203295] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.211600] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.304727] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.305298] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.308071] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.842s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.309598] env[61006]: INFO nova.compute.claims [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.389899] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336956, 'name': ReconfigVM_Task, 'duration_secs': 0.372076} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.390136] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Reconfigured VM instance instance-0000003e to attach disk [datastore2] ea7d5d49-ac76-4f2e-9456-912cf466fcc2/ea7d5d49-ac76-4f2e-9456-912cf466fcc2.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.390804] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8b18ae3-a65d-43fe-aca2-f0b4a22cbbe5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.397836] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 823.397836] env[61006]: value = "task-1336958" [ 823.397836] env[61006]: _type = "Task" [ 823.397836] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.408922] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336958, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.482060] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Successfully updated port: b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.644571] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.655923] env[61006]: INFO nova.compute.manager [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Took 29.49 seconds to build instance. [ 823.714886] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336957, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.815170] env[61006]: DEBUG nova.compute.utils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.820399] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.820399] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.859425] env[61006]: DEBUG nova.network.neutron [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updating instance_info_cache with network_info: [{"id": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "address": "fa:16:3e:f0:17:57", "network": {"id": "eba91a43-0fae-4db5-99f1-134ac43cea76", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1804121233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24ba46b1b91f46479841ea9db9dce16f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7043ca7a-807c-4c7b-b646-23ffece188b2", "external-id": "nsx-vlan-transportzone-619", "segmentation_id": 619, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0652e5ba-2f", "ovs_interfaceid": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.893921] env[61006]: DEBUG nova.policy [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42027a2dce454ca98e295dadca850eee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2b5a4bf20a40ee9b8688935de75dc1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.908839] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336958, 'name': Rename_Task, 'duration_secs': 0.438752} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.909283] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.909620] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b447c4fd-3152-45db-99f5-155966746492 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.917298] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 823.917298] env[61006]: value = "task-1336959" [ 823.917298] env[61006]: _type = "Task" [ 823.917298] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.928297] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.988020] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.988020] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.988020] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.020600] env[61006]: DEBUG nova.compute.manager [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Received event network-vif-plugged-b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.020808] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Acquiring lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.021125] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.021211] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.021464] env[61006]: DEBUG nova.compute.manager [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] No waiting events found dispatching network-vif-plugged-b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.021658] env[61006]: WARNING nova.compute.manager [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Received unexpected event network-vif-plugged-b2c4d5fa-8972-4475-af28-caa8b5ffce39 for instance with vm_state building and task_state spawning. [ 824.021847] env[61006]: DEBUG nova.compute.manager [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Received event network-changed-b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.022015] env[61006]: DEBUG nova.compute.manager [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Refreshing instance network info cache due to event network-changed-b2c4d5fa-8972-4475-af28-caa8b5ffce39. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.022295] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Acquiring lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.159370] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7d9170b6-2b2e-4735-9ec9-799d90e3e99c tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.983s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.184976] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Successfully created port: 608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.214872] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555249} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.215261] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 2a4089c9-4229-40bd-8d0e-706bba94655f/2a4089c9-4229-40bd-8d0e-706bba94655f.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.215525] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.215810] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d83e531c-df74-4797-a275-310b1fb2c083 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.222295] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 824.222295] env[61006]: value = "task-1336960" [ 824.222295] env[61006]: _type = "Task" [ 824.222295] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.230302] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336960, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.320045] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.362750] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Releasing lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.363070] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance network_info: |[{"id": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "address": "fa:16:3e:f0:17:57", "network": {"id": "eba91a43-0fae-4db5-99f1-134ac43cea76", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1804121233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24ba46b1b91f46479841ea9db9dce16f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7043ca7a-807c-4c7b-b646-23ffece188b2", "external-id": "nsx-vlan-transportzone-619", "segmentation_id": 619, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0652e5ba-2f", "ovs_interfaceid": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.363694] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:17:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7043ca7a-807c-4c7b-b646-23ffece188b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0652e5ba-2f85-49a7-85aa-2f0bc1757b4a', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.371664] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Creating folder: Project (24ba46b1b91f46479841ea9db9dce16f). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.373341] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93c01e9f-3acf-44c3-9316-9450bb32c674 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.386756] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Created folder: Project (24ba46b1b91f46479841ea9db9dce16f) in parent group-v285275. [ 824.386948] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Creating folder: Instances. Parent ref: group-v285317. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.389561] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfbdb133-46a8-4326-b705-b3dc1fcee01a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.398845] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Created folder: Instances in parent group-v285317. [ 824.399105] env[61006]: DEBUG oslo.service.loopingcall [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.399311] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.399520] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-027716dc-1c3e-443c-936d-634b0721b1f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.426882] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.426882] env[61006]: value = "task-1336963" [ 824.426882] env[61006]: _type = "Task" [ 824.426882] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.436599] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336959, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.442010] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336963, 'name': CreateVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.542770] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.660723] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.693972] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ef6e8d-a02b-4f50-b65b-335b327b17a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.702468] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a7b677-b7cd-4a98-b57f-9e653aa653c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.741237] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01450fa0-50cf-435e-97bd-45a99b841291 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.757065] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336960, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068327} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.759725] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.760619] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b9cc9b-a5dd-422a-9282-b9e41d0ac1f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.764928] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937f5fda-4d34-496c-9eb5-9f3e9bf58ef0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.787700] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 2a4089c9-4229-40bd-8d0e-706bba94655f/2a4089c9-4229-40bd-8d0e-706bba94655f.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.795661] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0318700d-cfcc-432c-8ce1-024d47dd234d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.809795] env[61006]: DEBUG nova.compute.provider_tree [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.815022] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Updating instance_info_cache with network_info: [{"id": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "address": "fa:16:3e:c4:01:39", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c4d5fa-89", "ovs_interfaceid": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.820141] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 824.820141] env[61006]: value = "task-1336964" [ 824.820141] env[61006]: _type = "Task" [ 824.820141] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.835940] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336964, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.931162] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336959, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.939426] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336963, 'name': CreateVM_Task, 'duration_secs': 0.374782} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.939604] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.940372] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.940542] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.940874] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 824.941180] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a24d75a2-7528-46d7-8b8c-4668f9476253 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.945978] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 824.945978] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528e2620-c749-e3aa-8859-1efb47bf8b98" [ 824.945978] env[61006]: _type = "Task" [ 824.945978] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.954538] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528e2620-c749-e3aa-8859-1efb47bf8b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.181559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.314591] env[61006]: DEBUG nova.scheduler.client.report [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.318670] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.318670] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Instance network_info: |[{"id": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "address": "fa:16:3e:c4:01:39", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c4d5fa-89", "ovs_interfaceid": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.318929] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Acquired lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.319130] env[61006]: DEBUG nova.network.neutron [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Refreshing network info cache for port b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.320348] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:01:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2c4d5fa-8972-4475-af28-caa8b5ffce39', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.327917] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating folder: Project (5d2b5a4bf20a40ee9b8688935de75dc1). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.331439] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a88ee67a-af33-420b-9cef-3a186d3fca43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.340384] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.351021] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336964, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.352763] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created folder: Project (5d2b5a4bf20a40ee9b8688935de75dc1) in parent group-v285275. [ 825.353022] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating folder: Instances. Parent ref: group-v285320. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.353320] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7864990f-1eba-4110-b71b-fd2dda61a60b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.363918] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.364242] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.364425] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.364664] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.364854] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.365024] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.365288] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.365492] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.365677] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.365888] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.366145] env[61006]: DEBUG nova.virt.hardware [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.367145] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff35758-959a-4fb5-a33a-0a2fc5429fe3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.373979] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created folder: Instances in parent group-v285320. [ 825.374314] env[61006]: DEBUG oslo.service.loopingcall [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.375074] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.376050] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c851a6d-35a6-47ae-b7e7-5822f161858f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.395511] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4fced1-ecd6-4178-ba29-1190eb166bee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.401330] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.401330] env[61006]: value = "task-1336967" [ 825.401330] env[61006]: _type = "Task" [ 825.401330] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.419582] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336967, 'name': CreateVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.430767] env[61006]: DEBUG oslo_vmware.api [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336959, 'name': PowerOnVM_Task, 'duration_secs': 1.356298} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.430767] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.430954] env[61006]: INFO nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Took 9.84 seconds to spawn the instance on the hypervisor. [ 825.431113] env[61006]: DEBUG nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.431978] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f0c477-e25d-437b-b40d-f3f24dd61fc3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.457673] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528e2620-c749-e3aa-8859-1efb47bf8b98, 'name': SearchDatastore_Task, 'duration_secs': 0.013003} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.457987] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.458244] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.458509] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.458686] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.458885] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.459367] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bddacae-e973-498f-b878-aefd75f7a243 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.472786] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.473025] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.473881] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e12f835d-97d6-4bb4-8b79-5583d6151a36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.480085] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 825.480085] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5256e081-eeb9-b04a-df86-0559b57524b7" [ 825.480085] env[61006]: _type = "Task" [ 825.480085] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.490479] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5256e081-eeb9-b04a-df86-0559b57524b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.655580] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "88e2bdc9-ab73-4e23-94b5-a45046835144" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.655959] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.656312] env[61006]: INFO nova.compute.manager [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Rebooting instance [ 825.675997] env[61006]: DEBUG nova.network.neutron [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Updated VIF entry in instance network info cache for port b2c4d5fa-8972-4475-af28-caa8b5ffce39. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.676409] env[61006]: DEBUG nova.network.neutron [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Updating instance_info_cache with network_info: [{"id": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "address": "fa:16:3e:c4:01:39", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2c4d5fa-89", "ovs_interfaceid": "b2c4d5fa-8972-4475-af28-caa8b5ffce39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.819825] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.820410] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.824429] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.313s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.825151] env[61006]: INFO nova.compute.claims [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.827993] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Successfully updated port: 608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.841726] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336964, 'name': ReconfigVM_Task, 'duration_secs': 0.576967} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.842273] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 2a4089c9-4229-40bd-8d0e-706bba94655f/2a4089c9-4229-40bd-8d0e-706bba94655f.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.843095] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53315060-aa86-4615-baae-44b10bca1bc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.852020] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 825.852020] env[61006]: value = "task-1336968" [ 825.852020] env[61006]: _type = "Task" [ 825.852020] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.860069] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336968, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.914129] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336967, 'name': CreateVM_Task, 'duration_secs': 0.382587} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.914296] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.915127] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.915341] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.915727] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.916045] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e460b5df-7542-4137-87c5-ce79c745ce43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.921499] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 825.921499] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d720d-a2cc-cde2-6587-ddad0d63e24a" [ 825.921499] env[61006]: _type = "Task" [ 825.921499] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.931077] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d720d-a2cc-cde2-6587-ddad0d63e24a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.951106] env[61006]: INFO nova.compute.manager [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Took 29.29 seconds to build instance. [ 825.994121] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5256e081-eeb9-b04a-df86-0559b57524b7, 'name': SearchDatastore_Task, 'duration_secs': 0.01371} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.995134] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35a60fd8-7483-47fa-a538-0ad24a62e4f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.001528] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 826.001528] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52aa2962-3ef1-6248-6319-3a3aadaa20bf" [ 826.001528] env[61006]: _type = "Task" [ 826.001528] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.010937] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52aa2962-3ef1-6248-6319-3a3aadaa20bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.075936] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.077049] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing instance network info cache due to event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.077049] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Acquiring lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.077049] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Acquired lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.077049] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 826.179528] env[61006]: DEBUG oslo_concurrency.lockutils [req-77c03b59-b33c-4b64-83a4-91817bb3c3a7 req-fdb986bf-9d2a-4018-87df-aef7324fd6b3 service nova] Releasing lock "refresh_cache-598c0d72-d679-49a8-b17c-f5f341c205e8" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.183186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.329638] env[61006]: DEBUG nova.compute.utils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.333124] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.333261] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.333403] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.334382] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.334549] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.359653] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336968, 'name': Rename_Task, 'duration_secs': 0.208179} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.359913] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.360184] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d10be0c-35db-4dfc-9733-5f324c4cc79e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.366839] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 826.366839] env[61006]: value = "task-1336969" [ 826.366839] env[61006]: _type = "Task" [ 826.366839] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.374594] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.402888] env[61006]: DEBUG nova.policy [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2867fd5b65d743e2a538f73855de1859', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c28711bb635749fcbf46eb8b43327fec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.433225] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524d720d-a2cc-cde2-6587-ddad0d63e24a, 'name': SearchDatastore_Task, 'duration_secs': 0.017509} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.433579] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.433818] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.434050] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.457072] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bd790830-0107-41a2-8fe2-371ac830a56b tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.569s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.515223] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52aa2962-3ef1-6248-6319-3a3aadaa20bf, 'name': SearchDatastore_Task, 'duration_secs': 0.027311} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.515886] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.516216] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 42b92d52-d1f0-48ff-94b6-6164b387456c/42b92d52-d1f0-48ff-94b6-6164b387456c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.516784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.516982] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.517288] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a1305f9-d74a-459f-856c-9edb69fea426 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.519609] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cfa6407-bc35-4a0f-a36d-b292246f6189 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.526652] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 826.526652] env[61006]: value = "task-1336970" [ 826.526652] env[61006]: _type = "Task" [ 826.526652] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.531579] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.531936] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.533108] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdca0e2a-0404-4b44-b3a3-1063a67120f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.539795] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.544018] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 826.544018] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52124f51-7096-5832-5ec0-f4d7a0cc3971" [ 826.544018] env[61006]: _type = "Task" [ 826.544018] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.551156] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52124f51-7096-5832-5ec0-f4d7a0cc3971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.781499] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.782628] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.783014] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.783363] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.783682] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.786955] env[61006]: INFO nova.compute.manager [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Terminating instance [ 826.791836] env[61006]: DEBUG nova.compute.manager [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.792232] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.793473] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e83b623-54d6-48a5-b678-def9d060010e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.803224] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 826.803224] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ccc7949-e2ba-471b-a687-99e0f5e9a982 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.809101] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 826.809101] env[61006]: value = "task-1336971" [ 826.809101] env[61006]: _type = "Task" [ 826.809101] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.818346] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.821558] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Successfully created port: 52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.845032] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.888960] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.911614] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.959519] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.043362] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336970, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.054484] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52124f51-7096-5832-5ec0-f4d7a0cc3971, 'name': SearchDatastore_Task, 'duration_secs': 0.015191} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.058837] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07b2190b-7d60-4e88-a743-5d76734578f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.065187] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 827.065187] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c704-c21c-18ea-d047-4f3a2868ee10" [ 827.065187] env[61006]: _type = "Task" [ 827.065187] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.073950] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c704-c21c-18ea-d047-4f3a2868ee10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.197236] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updated VIF entry in instance network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.198246] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.254429] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7497eec8-a927-49be-999b-e81e54b4dbf7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.263048] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1753be1a-8a14-4a81-be28-9a9c9bd33bfc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.310836] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd06f199-2cab-410a-9181-470623ef3c2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.328275] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10393684-8986-4854-b2ff-5f818684544b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.333840] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336971, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.349924] env[61006]: DEBUG nova.compute.provider_tree [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.378674] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.420734] env[61006]: DEBUG nova.network.neutron [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Updating instance_info_cache with network_info: [{"id": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "address": "fa:16:3e:d7:1f:e9", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608225ce-37", "ovs_interfaceid": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.482319] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.538317] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710964} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.538619] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 42b92d52-d1f0-48ff-94b6-6164b387456c/42b92d52-d1f0-48ff-94b6-6164b387456c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.539615] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.539615] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc7e2252-1765-4869-bb98-f0db75497130 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.546251] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 827.546251] env[61006]: value = "task-1336972" [ 827.546251] env[61006]: _type = "Task" [ 827.546251] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.557642] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.575547] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0c704-c21c-18ea-d047-4f3a2868ee10, 'name': SearchDatastore_Task, 'duration_secs': 0.062765} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.575547] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.575708] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 598c0d72-d679-49a8-b17c-f5f341c205e8/598c0d72-d679-49a8-b17c-f5f341c205e8.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.575912] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-877bbfd3-14ca-4359-a326-36d00e9c9932 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.583679] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 827.583679] env[61006]: value = "task-1336973" [ 827.583679] env[61006]: _type = "Task" [ 827.583679] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.707695] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Releasing lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.708516] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Received event network-vif-plugged-608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.708516] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Acquiring lock "29c41817-2189-4622-8f35-86f61eb34bed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.708516] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Lock "29c41817-2189-4622-8f35-86f61eb34bed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.708704] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Lock "29c41817-2189-4622-8f35-86f61eb34bed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.708789] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] No waiting events found dispatching network-vif-plugged-608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.708934] env[61006]: WARNING nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Received unexpected event network-vif-plugged-608225ce-37a4-4c41-ba85-f2e0575b97d1 for instance with vm_state building and task_state spawning. [ 827.709215] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Received event network-changed-608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.709814] env[61006]: DEBUG nova.compute.manager [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Refreshing instance network info cache due to event network-changed-608225ce-37a4-4c41-ba85-f2e0575b97d1. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.709814] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Acquiring lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.709814] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquired lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.710048] env[61006]: DEBUG nova.network.neutron [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.827108] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336971, 'name': PowerOffVM_Task, 'duration_secs': 0.533171} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.827468] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.827695] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.827968] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9b21fd6-3f9c-4a55-8e27-bd9b5d90a8de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.852874] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.856052] env[61006]: DEBUG nova.scheduler.client.report [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.881393] env[61006]: DEBUG oslo_vmware.api [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336969, 'name': PowerOnVM_Task, 'duration_secs': 1.460292} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.883860] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.884172] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.884404] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.884609] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.884776] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.884938] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.885202] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.885398] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.885671] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.885927] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.886160] env[61006]: DEBUG nova.virt.hardware [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.886493] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.886718] env[61006]: INFO nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Took 9.86 seconds to spawn the instance on the hypervisor. [ 827.886930] env[61006]: DEBUG nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.888200] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8220a7cb-910e-4b91-b989-58a908b5706b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.892565] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9f436f-46db-428b-b9f1-025fc8e6b853 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.895716] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.895930] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.896150] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Deleting the datastore file [datastore2] ea7d5d49-ac76-4f2e-9456-912cf466fcc2 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.897098] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fd665d8-9cf7-4665-89b1-db0606e13e3e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.904199] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41cd600-677f-470f-9eee-8f65d23d4acb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.925025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.925306] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Instance network_info: |[{"id": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "address": "fa:16:3e:d7:1f:e9", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608225ce-37", "ovs_interfaceid": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.927045] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Acquired lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.927277] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Refreshing network info cache for port 608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.928571] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:1f:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '608225ce-37a4-4c41-ba85-f2e0575b97d1', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.936437] env[61006]: DEBUG oslo.service.loopingcall [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.936794] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for the task: (returnval){ [ 827.936794] env[61006]: value = "task-1336975" [ 827.936794] env[61006]: _type = "Task" [ 827.936794] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.937871] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.938187] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d471abcb-b85c-4ac7-8723-d977aaa0a3bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.965420] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336975, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.967066] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.967066] env[61006]: value = "task-1336976" [ 827.967066] env[61006]: _type = "Task" [ 827.967066] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.977468] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336976, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.057254] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076146} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.058386] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.058660] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c68a592-a3e3-4f71-9123-9d819fd36854 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.083842] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 42b92d52-d1f0-48ff-94b6-6164b387456c/42b92d52-d1f0-48ff-94b6-6164b387456c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.084473] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4469d7e-9208-47d2-b343-a72c1f2caacb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.108865] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336973, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.110450] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 828.110450] env[61006]: value = "task-1336977" [ 828.110450] env[61006]: _type = "Task" [ 828.110450] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.120318] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.361619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.361940] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.367656] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.491s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.369186] env[61006]: INFO nova.compute.claims [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.428634] env[61006]: INFO nova.compute.manager [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Took 29.72 seconds to build instance. [ 828.450391] env[61006]: DEBUG oslo_vmware.api [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Task: {'id': task-1336975, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.521744} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.450828] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.451145] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.451480] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.451779] env[61006]: INFO nova.compute.manager [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Took 1.66 seconds to destroy the instance on the hypervisor. [ 828.452165] env[61006]: DEBUG oslo.service.loopingcall [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.452469] env[61006]: DEBUG nova.compute.manager [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.452663] env[61006]: DEBUG nova.network.neutron [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.476754] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336976, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.508654] env[61006]: DEBUG nova.network.neutron [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.598023] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684197} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.598023] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 598c0d72-d679-49a8-b17c-f5f341c205e8/598c0d72-d679-49a8-b17c-f5f341c205e8.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.598023] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.598023] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5303c540-08bd-4560-9574-f0b55ea89796 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.605237] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 828.605237] env[61006]: value = "task-1336978" [ 828.605237] env[61006]: _type = "Task" [ 828.605237] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.620777] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.630621] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336977, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.874682] env[61006]: DEBUG nova.compute.utils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.878275] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.878447] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 828.890556] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Updated VIF entry in instance network info cache for port 608225ce-37a4-4c41-ba85-f2e0575b97d1. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.891408] env[61006]: DEBUG nova.network.neutron [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Updating instance_info_cache with network_info: [{"id": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "address": "fa:16:3e:d7:1f:e9", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap608225ce-37", "ovs_interfaceid": "608225ce-37a4-4c41-ba85-f2e0575b97d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.930936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c33352a9-79db-483b-bc93-72fc5b3e8c68 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.417s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.959046] env[61006]: DEBUG nova.policy [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7707895279cb4c6ebc07bf3570f3f881', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c673bf1b8a437fbfbfd34e912a8f37', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.981615] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336976, 'name': CreateVM_Task, 'duration_secs': 0.551749} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.981745] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.982557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.982975] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.983252] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.983563] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb03ace0-711a-4aa2-a0b3-02939fdee436 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.989089] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 828.989089] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c48c53-e30d-4381-1c88-8720fab03a1b" [ 828.989089] env[61006]: _type = "Task" [ 828.989089] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.998403] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c48c53-e30d-4381-1c88-8720fab03a1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.011801] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Releasing lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.014834] env[61006]: DEBUG nova.compute.manager [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.015762] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2c48c2-9c66-4320-806e-1bd5692aed68 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.054478] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "2a4089c9-4229-40bd-8d0e-706bba94655f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.054787] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.056017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.056310] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.056656] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.063306] env[61006]: INFO nova.compute.manager [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Terminating instance [ 829.065927] env[61006]: DEBUG nova.compute.manager [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.066729] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.068382] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c63a45-2bd3-4090-8eb3-00334e9336d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.078969] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.079374] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f035b089-6e53-4a00-b919-3ff1ae9d28c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.085623] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 829.085623] env[61006]: value = "task-1336979" [ 829.085623] env[61006]: _type = "Task" [ 829.085623] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.098636] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.099473] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Successfully updated port: 52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.120388] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08014} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.121176] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.122047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd20f178-fe61-4101-a563-89df253e8b68 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.129220] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336977, 'name': ReconfigVM_Task, 'duration_secs': 0.522683} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.129940] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 42b92d52-d1f0-48ff-94b6-6164b387456c/42b92d52-d1f0-48ff-94b6-6164b387456c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.130588] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e462352-1875-4668-b4a8-5f14c28cb06d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.158348] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 598c0d72-d679-49a8-b17c-f5f341c205e8/598c0d72-d679-49a8-b17c-f5f341c205e8.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.158348] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-137c7f56-f903-468e-ad6f-dd2adc586cf6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.181015] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 829.181015] env[61006]: value = "task-1336980" [ 829.181015] env[61006]: _type = "Task" [ 829.181015] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.187389] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 829.187389] env[61006]: value = "task-1336981" [ 829.187389] env[61006]: _type = "Task" [ 829.187389] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.200538] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336980, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.207875] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336981, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.259468] env[61006]: DEBUG nova.compute.manager [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Received event network-vif-plugged-52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.259713] env[61006]: DEBUG oslo_concurrency.lockutils [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] Acquiring lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.259918] env[61006]: DEBUG oslo_concurrency.lockutils [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.263649] env[61006]: DEBUG oslo_concurrency.lockutils [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.263649] env[61006]: DEBUG nova.compute.manager [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] No waiting events found dispatching network-vif-plugged-52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.263649] env[61006]: WARNING nova.compute.manager [req-1be61148-a7bf-4d1a-a682-ac95ac9c6284 req-bf7f9167-5d97-4a41-ab24-46656c0a2883 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Received unexpected event network-vif-plugged-52eb94ea-4588-467f-9b0c-273dbbae4f8b for instance with vm_state building and task_state spawning. [ 829.263649] env[61006]: DEBUG nova.compute.manager [req-125039e2-bbf6-416e-ada2-e91800d863d7 req-505b7c8b-88fb-46c8-8a3a-12881314fd12 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Received event network-vif-deleted-8dbbd99a-28ee-4db1-bf70-63d8827606c2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.263649] env[61006]: INFO nova.compute.manager [req-125039e2-bbf6-416e-ada2-e91800d863d7 req-505b7c8b-88fb-46c8-8a3a-12881314fd12 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Neutron deleted interface 8dbbd99a-28ee-4db1-bf70-63d8827606c2; detaching it from the instance and deleting it from the info cache [ 829.264109] env[61006]: DEBUG nova.network.neutron [req-125039e2-bbf6-416e-ada2-e91800d863d7 req-505b7c8b-88fb-46c8-8a3a-12881314fd12 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.348646] env[61006]: DEBUG nova.network.neutron [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.381735] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.393758] env[61006]: DEBUG oslo_concurrency.lockutils [req-306fac3c-ec49-4de2-a1bd-167bfeaf9208 req-5446ae42-edab-4421-9b90-4765101f7097 service nova] Releasing lock "refresh_cache-29c41817-2189-4622-8f35-86f61eb34bed" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.398764] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Successfully created port: 3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.436043] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.501251] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c48c53-e30d-4381-1c88-8720fab03a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.011722} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.501674] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.501919] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.502603] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.502603] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.502603] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.502774] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3940b26-813a-4aad-a7a5-cb2620c5042a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.514625] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.515691] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.515691] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bf4f120-ba82-4e4a-9698-d0f3df113146 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.522304] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 829.522304] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527435c9-3fa7-5c1a-2873-661176192e10" [ 829.522304] env[61006]: _type = "Task" [ 829.522304] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.539867] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527435c9-3fa7-5c1a-2873-661176192e10, 'name': SearchDatastore_Task, 'duration_secs': 0.011053} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.539867] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60374747-c605-4d9a-9c94-25dfac694b12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.547707] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 829.547707] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52977569-5f11-abac-2c70-ba68390ff7e6" [ 829.547707] env[61006]: _type = "Task" [ 829.547707] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.556102] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52977569-5f11-abac-2c70-ba68390ff7e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.599013] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336979, 'name': PowerOffVM_Task, 'duration_secs': 0.277516} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.599237] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.599449] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.599803] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e318419e-07e3-4eb8-8ee9-d5ec84c322f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.606345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.606345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.606345] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.657656] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 829.658627] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f632896-3b45-4cb0-8c55-34cb75c47a27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.668072] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 829.668351] env[61006]: ERROR oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk due to incomplete transfer. [ 829.668605] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8ee12a56-b6c8-4502-8edd-d4f0a13cfaf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.677359] env[61006]: DEBUG oslo_vmware.rw_handles [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5219d898-8c1e-2643-c63d-091219e5ca80/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 829.677631] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Uploaded image 06d8ddf9-1620-4813-97cd-7b51789db884 to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 829.679907] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 829.681154] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-580038ec-3d29-4607-a605-301ef6ea613b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.682808] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.683019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.683209] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Deleting the datastore file [datastore2] 2a4089c9-4229-40bd-8d0e-706bba94655f {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.683512] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80b51a3d-da02-4bf1-af6d-8ede39503aca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.695994] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 829.695994] env[61006]: value = "task-1336983" [ 829.695994] env[61006]: _type = "Task" [ 829.695994] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.700239] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336980, 'name': Rename_Task, 'duration_secs': 0.286464} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.700571] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for the task: (returnval){ [ 829.700571] env[61006]: value = "task-1336984" [ 829.700571] env[61006]: _type = "Task" [ 829.700571] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.704201] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.709576] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f2e1256-e863-4102-9aad-ee57d5d077bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.715705] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336981, 'name': ReconfigVM_Task, 'duration_secs': 0.492007} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.716517] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 598c0d72-d679-49a8-b17c-f5f341c205e8/598c0d72-d679-49a8-b17c-f5f341c205e8.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.717644] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9762b9ce-4932-4c2b-97f2-1098e0c8fb9b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.725461] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.725720] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336983, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.727887] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 829.727887] env[61006]: value = "task-1336985" [ 829.727887] env[61006]: _type = "Task" [ 829.727887] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.740140] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336985, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.741438] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 829.741438] env[61006]: value = "task-1336986" [ 829.741438] env[61006]: _type = "Task" [ 829.741438] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.749282] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336986, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.767319] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-901f15e3-39c1-4472-b589-e33050b7bc54 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.776450] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4a197f-9803-4cc9-8051-51b79b1b3a5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.806639] env[61006]: DEBUG nova.compute.manager [req-125039e2-bbf6-416e-ada2-e91800d863d7 req-505b7c8b-88fb-46c8-8a3a-12881314fd12 service nova] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Detach interface failed, port_id=8dbbd99a-28ee-4db1-bf70-63d8827606c2, reason: Instance ea7d5d49-ac76-4f2e-9456-912cf466fcc2 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 829.808583] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abfa777-b11c-444a-8ff2-6f4ceab1aaed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.815949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b349bc64-1ee9-4e98-b532-707b6c9cd012 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.847360] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d3a80b-dec2-4714-989f-0ae40439197b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.851864] env[61006]: INFO nova.compute.manager [-] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Took 1.40 seconds to deallocate network for instance. [ 829.856613] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282f9d89-d1a3-420d-9996-8811b22420f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.874148] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 829.954092] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.036625] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6186fb3-57cb-4b90-a6de-33467458d0bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.044425] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Doing hard reboot of VM {{(pid=61006) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 830.044672] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0f53ba14-887f-4e23-9f61-f24dc0844b1b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.052720] env[61006]: DEBUG oslo_vmware.api [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 830.052720] env[61006]: value = "task-1336987" [ 830.052720] env[61006]: _type = "Task" [ 830.052720] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.059778] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52977569-5f11-abac-2c70-ba68390ff7e6, 'name': SearchDatastore_Task, 'duration_secs': 0.011152} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.060462] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.060787] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 29c41817-2189-4622-8f35-86f61eb34bed/29c41817-2189-4622-8f35-86f61eb34bed.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.061075] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4fdf2c9a-11a5-486f-8183-2ec33d725284 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.066201] env[61006]: DEBUG oslo_vmware.api [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336987, 'name': ResetVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.070668] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 830.070668] env[61006]: value = "task-1336988" [ 830.070668] env[61006]: _type = "Task" [ 830.070668] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.078542] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.142656] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.215253] env[61006]: DEBUG oslo_vmware.api [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Task: {'id': task-1336984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182334} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.218569] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.218793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 830.218997] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.219217] env[61006]: INFO nova.compute.manager [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 830.219508] env[61006]: DEBUG oslo.service.loopingcall [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.219727] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336983, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.219955] env[61006]: DEBUG nova.compute.manager [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.220115] env[61006]: DEBUG nova.network.neutron [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 830.239571] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336985, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.250378] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336986, 'name': Rename_Task, 'duration_secs': 0.259205} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.250646] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.250887] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3b22d8d-d89e-4164-b2e4-cdadbdd392cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.256941] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 830.256941] env[61006]: value = "task-1336989" [ 830.256941] env[61006]: _type = "Task" [ 830.256941] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.266388] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.365953] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.395293] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.406041] env[61006]: ERROR nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [req-62873890-67ed-4849-bf0a-ad3f2f94a2dc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3360045e-46ab-4f2d-9377-dd481ab3cd53. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-62873890-67ed-4849-bf0a-ad3f2f94a2dc"}]} [ 830.414946] env[61006]: DEBUG nova.network.neutron [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Updating instance_info_cache with network_info: [{"id": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "address": "fa:16:3e:34:8f:d0", "network": {"id": "198771be-fdd5-4687-9f69-07c254a92200", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-358749898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28711bb635749fcbf46eb8b43327fec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52eb94ea-45", "ovs_interfaceid": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.420825] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.421122] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.421283] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.421553] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.421702] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.421848] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.422067] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.422226] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.422411] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.422664] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.422869] env[61006]: DEBUG nova.virt.hardware [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.423822] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1162e67b-fbaa-406a-a28d-14c28bffd3a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.427865] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing inventories for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 830.436866] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f766f6f4-561a-4cac-b42b-e45aca432e61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.453772] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating ProviderTree inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 830.453983] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.470740] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing aggregate associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, aggregates: None {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 830.496084] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing trait associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 830.565451] env[61006]: DEBUG oslo_vmware.api [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336987, 'name': ResetVM_Task, 'duration_secs': 0.097565} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.568548] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Did hard reboot of VM {{(pid=61006) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 830.568844] env[61006]: DEBUG nova.compute.manager [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.570544] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8c94d9-05e4-4d77-8fd7-ffd8678a333d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.583784] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336988, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.712747] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336983, 'name': Destroy_Task, 'duration_secs': 0.725154} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.713041] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Destroyed the VM [ 830.713369] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 830.713627] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7967c5df-45b3-46fb-9a80-62d904088080 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.719949] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 830.719949] env[61006]: value = "task-1336990" [ 830.719949] env[61006]: _type = "Task" [ 830.719949] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.727652] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336990, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.738578] env[61006]: DEBUG oslo_vmware.api [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1336985, 'name': PowerOnVM_Task, 'duration_secs': 0.869628} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.738865] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.739091] env[61006]: INFO nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Took 10.30 seconds to spawn the instance on the hypervisor. [ 830.739279] env[61006]: DEBUG nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.740043] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0916a2-e86a-44f9-8515-3fc25397aee8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.767073] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336989, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.835768] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2476e1-e5d2-4d9c-b4c0-6bd7017ec65d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.843740] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fc24d1-64c7-4535-8f3a-f2b6de70a80e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.873382] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0906a836-e851-42d1-8e8d-3adfc8611861 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.880812] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da3fd3c-4478-4bf8-8532-8a60ba8683ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.895243] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.917403] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Releasing lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.917651] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Instance network_info: |[{"id": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "address": "fa:16:3e:34:8f:d0", "network": {"id": "198771be-fdd5-4687-9f69-07c254a92200", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-358749898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28711bb635749fcbf46eb8b43327fec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52eb94ea-45", "ovs_interfaceid": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.918077] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:8f:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52eb94ea-4588-467f-9b0c-273dbbae4f8b', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.926343] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Creating folder: Project (c28711bb635749fcbf46eb8b43327fec). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.926343] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-188fa72e-ef8c-4d40-8b68-c6d0cc4e9152 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.935596] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Created folder: Project (c28711bb635749fcbf46eb8b43327fec) in parent group-v285275. [ 830.935800] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Creating folder: Instances. Parent ref: group-v285324. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.936027] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a830c3d9-361a-4b82-bf4d-5fb86babc96a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.944788] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Created folder: Instances in parent group-v285324. [ 830.945113] env[61006]: DEBUG oslo.service.loopingcall [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.945380] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.945670] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58ba4f98-d5d9-432b-9544-0231b662b39f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.974586] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.974586] env[61006]: value = "task-1336993" [ 830.974586] env[61006]: _type = "Task" [ 830.974586] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.985661] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336993, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.082920] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572698} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.083207] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 29c41817-2189-4622-8f35-86f61eb34bed/29c41817-2189-4622-8f35-86f61eb34bed.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.083454] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.083716] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0475d08-798b-44b2-8bd3-798fd47d9809 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.089929] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 831.089929] env[61006]: value = "task-1336994" [ 831.089929] env[61006]: _type = "Task" [ 831.089929] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.096078] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5d9f18e0-9a04-47a8-94e3-e16e7ceeb869 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.440s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.100380] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.169638] env[61006]: DEBUG nova.network.neutron [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.230771] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336990, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.272689] env[61006]: INFO nova.compute.manager [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Took 31.01 seconds to build instance. [ 831.280408] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336989, 'name': PowerOnVM_Task, 'duration_secs': 0.913508} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.280724] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 831.280906] env[61006]: INFO nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Took 8.41 seconds to spawn the instance on the hypervisor. [ 831.281103] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.282040] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35660c12-4daa-4c85-b80d-259939177be0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.301561] env[61006]: DEBUG nova.compute.manager [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Received event network-vif-plugged-3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.301772] env[61006]: DEBUG oslo_concurrency.lockutils [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.301828] env[61006]: DEBUG oslo_concurrency.lockutils [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.302153] env[61006]: DEBUG oslo_concurrency.lockutils [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.302488] env[61006]: DEBUG nova.compute.manager [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] No waiting events found dispatching network-vif-plugged-3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.302527] env[61006]: WARNING nova.compute.manager [req-ee3387f4-7b46-462e-a87a-25544dcd82f2 req-0f443fce-3c66-449d-8a07-abdd2d4a0340 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Received unexpected event network-vif-plugged-3d9ba3b1-a27e-4002-bd16-cfd881fca452 for instance with vm_state building and task_state spawning. [ 831.304919] env[61006]: DEBUG nova.compute.manager [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Received event network-changed-52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.305127] env[61006]: DEBUG nova.compute.manager [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Refreshing instance network info cache due to event network-changed-52eb94ea-4588-467f-9b0c-273dbbae4f8b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.305765] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] Acquiring lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.305765] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] Acquired lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.305765] env[61006]: DEBUG nova.network.neutron [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Refreshing network info cache for port 52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.397546] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Successfully updated port: 3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.429288] env[61006]: ERROR nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [req-8008f309-c0c8-4b84-92ae-d8dac16b531e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3360045e-46ab-4f2d-9377-dd481ab3cd53. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8008f309-c0c8-4b84-92ae-d8dac16b531e"}]} [ 831.444891] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing inventories for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 831.456953] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating ProviderTree inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 831.457205] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.468162] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing aggregate associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, aggregates: None {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 831.486774] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336993, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.487834] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Refreshing trait associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 831.599838] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0795} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.602303] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.603470] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b93f4d-3f9c-4031-867f-11294e465597 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.625837] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 29c41817-2189-4622-8f35-86f61eb34bed/29c41817-2189-4622-8f35-86f61eb34bed.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.628337] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c204cdaf-6ca9-4ac4-8575-29ad8005de04 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.647521] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 831.647521] env[61006]: value = "task-1336995" [ 831.647521] env[61006]: _type = "Task" [ 831.647521] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.656934] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336995, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.672481] env[61006]: INFO nova.compute.manager [-] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Took 1.45 seconds to deallocate network for instance. [ 831.732067] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336990, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.775785] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9236b055-9e72-45f4-80f9-bdcf16f53826 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.557s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.806798] env[61006]: INFO nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Took 29.30 seconds to build instance. [ 831.832362] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "2232c77d-9f4e-4981-9f55-614bba5d71db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.832688] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.833482] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e130e9-b66e-4d3b-b4e3-bb7f0a3b70d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.842313] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981a7344-56dc-4e22-bd16-7538188961b5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.878281] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0772d8c-e5af-4721-8628-fff4eb92b154 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.887106] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf7533c-4b1f-4c18-8cbe-41198fd3862b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.902394] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.903581] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.903581] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.903581] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.987950] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1336993, 'name': CreateVM_Task, 'duration_secs': 0.82868} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.988159] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.988862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.989120] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.989368] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.989997] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68bc94b2-60de-49a9-9317-c6e879dc9092 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.995214] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 831.995214] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526d9e1b-6362-0284-b8f3-71e800abd5af" [ 831.995214] env[61006]: _type = "Task" [ 831.995214] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.006250] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526d9e1b-6362-0284-b8f3-71e800abd5af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.157841] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336995, 'name': ReconfigVM_Task, 'duration_secs': 0.464505} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.158139] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 29c41817-2189-4622-8f35-86f61eb34bed/29c41817-2189-4622-8f35-86f61eb34bed.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.159272] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7f4b815-38d5-445f-95b6-b7aabaad237e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.166548] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 832.166548] env[61006]: value = "task-1336996" [ 832.166548] env[61006]: _type = "Task" [ 832.166548] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.175472] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336996, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.181207] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.235090] env[61006]: DEBUG oslo_vmware.api [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1336990, 'name': RemoveSnapshot_Task, 'duration_secs': 1.329239} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.236072] env[61006]: DEBUG nova.network.neutron [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Updated VIF entry in instance network info cache for port 52eb94ea-4588-467f-9b0c-273dbbae4f8b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.236408] env[61006]: DEBUG nova.network.neutron [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Updating instance_info_cache with network_info: [{"id": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "address": "fa:16:3e:34:8f:d0", "network": {"id": "198771be-fdd5-4687-9f69-07c254a92200", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-358749898-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c28711bb635749fcbf46eb8b43327fec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52eb94ea-45", "ovs_interfaceid": "52eb94ea-4588-467f-9b0c-273dbbae4f8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.237605] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 832.237844] env[61006]: INFO nova.compute.manager [None req-4e5d1cf9-5891-43a0-826a-482ff669249d tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Took 15.21 seconds to snapshot the instance on the hypervisor. [ 832.278185] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.311932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.604s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.437120] env[61006]: DEBUG nova.scheduler.client.report [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 832.437429] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 85 to 86 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 832.440019] env[61006]: DEBUG nova.compute.provider_tree [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.464984] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.507934] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526d9e1b-6362-0284-b8f3-71e800abd5af, 'name': SearchDatastore_Task, 'duration_secs': 0.011064} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.508164] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.508432] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.508690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.508840] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.509027] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.509307] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c97c0c5-83a5-47a5-a216-2c59d966a69e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.517479] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.517712] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.520800] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9368fd02-8266-440b-95ee-9805d7916389 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.527127] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 832.527127] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd901-300c-84d1-5192-c50ecee55716" [ 832.527127] env[61006]: _type = "Task" [ 832.527127] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.537768] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd901-300c-84d1-5192-c50ecee55716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.682028] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336996, 'name': Rename_Task, 'duration_secs': 0.365704} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.682028] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.682028] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1eafd31-d892-4907-99f0-307f9f050d28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.686018] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 832.686018] env[61006]: value = "task-1336997" [ 832.686018] env[61006]: _type = "Task" [ 832.686018] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.692691] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.696231] env[61006]: DEBUG nova.network.neutron [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.744020] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] Releasing lock "refresh_cache-72c3e37b-2eac-41d4-8308-0a6466c2dd24" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.744020] env[61006]: DEBUG nova.compute.manager [req-b0aed037-b8ea-409a-bef5-416ef8b570e8 req-d5c00fc3-32f3-4f54-b6ab-e175b198faf1 service nova] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Received event network-vif-deleted-a94adb0b-5ee2-4af8-bfcf-f268d54f1eda {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.806125] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.817123] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.859065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "88e2bdc9-ab73-4e23-94b5-a45046835144" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.859065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.859065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.859065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.859339] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.860118] env[61006]: INFO nova.compute.manager [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Terminating instance [ 832.862736] env[61006]: DEBUG nova.compute.manager [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.863072] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.864148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b472867-08de-47da-8319-d8c3ea5efc75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.873080] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.873473] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccd68e41-8892-4f8e-a1c3-d32074049d0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.882037] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 832.882037] env[61006]: value = "task-1336998" [ 832.882037] env[61006]: _type = "Task" [ 832.882037] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.888128] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.947240] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.577s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.947240] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.948462] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.038s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.950045] env[61006]: INFO nova.compute.claims [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.039015] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd901-300c-84d1-5192-c50ecee55716, 'name': SearchDatastore_Task, 'duration_secs': 0.010025} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.039551] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf9ed1bb-3ee5-47ab-a874-988a6ee44b0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.046628] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 833.046628] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be54f0-c7bf-0f1d-aa13-9e7b7915a499" [ 833.046628] env[61006]: _type = "Task" [ 833.046628] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.054857] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be54f0-c7bf-0f1d-aa13-9e7b7915a499, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.195016] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336997, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.198608] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.198914] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Instance network_info: |[{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.199370] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:9c:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d9ba3b1-a27e-4002-bd16-cfd881fca452', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.206877] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating folder: Project (08c673bf1b8a437fbfbfd34e912a8f37). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.207174] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64246695-5ea9-4d65-ae02-40f42a22218e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.217598] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created folder: Project (08c673bf1b8a437fbfbfd34e912a8f37) in parent group-v285275. [ 833.217874] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating folder: Instances. Parent ref: group-v285327. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.218147] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a9885ca-bbde-4e33-a15a-d9d5b91049f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.227767] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created folder: Instances in parent group-v285327. [ 833.228058] env[61006]: DEBUG oslo.service.loopingcall [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.228294] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.228588] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40ac2d66-9b71-48ce-ad46-0df759449789 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.249084] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.249084] env[61006]: value = "task-1337001" [ 833.249084] env[61006]: _type = "Task" [ 833.249084] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.256825] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337001, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.338448] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.388424] env[61006]: DEBUG nova.compute.manager [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.388537] env[61006]: DEBUG nova.compute.manager [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing instance network info cache due to event network-changed-e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.388880] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Acquiring lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.389084] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Acquired lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.389243] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Refreshing network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.398327] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1336998, 'name': PowerOffVM_Task, 'duration_secs': 0.179933} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.399147] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.399333] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.399796] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4afa751-a6ef-4c1e-a60a-daf0ee10ba98 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.409347] env[61006]: DEBUG nova.compute.manager [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Received event network-changed-3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.409347] env[61006]: DEBUG nova.compute.manager [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Refreshing instance network info cache due to event network-changed-3d9ba3b1-a27e-4002-bd16-cfd881fca452. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.409347] env[61006]: DEBUG oslo_concurrency.lockutils [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.409347] env[61006]: DEBUG oslo_concurrency.lockutils [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.409347] env[61006]: DEBUG nova.network.neutron [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Refreshing network info cache for port 3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.456264] env[61006]: DEBUG nova.compute.utils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.460187] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.461243] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.467577] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.467805] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.468075] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Deleting the datastore file [datastore2] 88e2bdc9-ab73-4e23-94b5-a45046835144 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.470027] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c28ef22-cc4b-4d1c-af53-00769063cf8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.478862] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for the task: (returnval){ [ 833.478862] env[61006]: value = "task-1337003" [ 833.478862] env[61006]: _type = "Task" [ 833.478862] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.488125] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1337003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.510223] env[61006]: DEBUG nova.policy [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78136eea79b54005bb35e4cfc5acc3d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0df410358ec405088642ab7614b2293', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.561644] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be54f0-c7bf-0f1d-aa13-9e7b7915a499, 'name': SearchDatastore_Task, 'duration_secs': 0.012277} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.561970] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.562271] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 72c3e37b-2eac-41d4-8308-0a6466c2dd24/72c3e37b-2eac-41d4-8308-0a6466c2dd24.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.562587] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c80c6f2c-b4af-4269-ae43-5e2e7af7f65e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.570590] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 833.570590] env[61006]: value = "task-1337004" [ 833.570590] env[61006]: _type = "Task" [ 833.570590] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.579637] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.698979] env[61006]: DEBUG oslo_vmware.api [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1336997, 'name': PowerOnVM_Task, 'duration_secs': 0.716347} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.699322] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.699572] env[61006]: INFO nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Took 8.36 seconds to spawn the instance on the hypervisor. [ 833.699783] env[61006]: DEBUG nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.704644] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aed50b4-3e32-475e-8246-0cfb7b6191ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.760574] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337001, 'name': CreateVM_Task, 'duration_secs': 0.332368} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.760789] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.761585] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.761738] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.762122] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.762501] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c40d7162-1824-42ee-bbde-3ca6b33246c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.768403] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 833.768403] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ce1547-4ee0-15ab-75b9-45361eed0dc6" [ 833.768403] env[61006]: _type = "Task" [ 833.768403] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.777965] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ce1547-4ee0-15ab-75b9-45361eed0dc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.876387] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Successfully created port: b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.964705] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.987704] env[61006]: DEBUG oslo_vmware.api [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Task: {'id': task-1337003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139617} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.988175] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.988593] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.988691] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.988918] env[61006]: INFO nova.compute.manager [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Took 1.13 seconds to destroy the instance on the hypervisor. [ 833.989590] env[61006]: DEBUG oslo.service.loopingcall [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.989978] env[61006]: DEBUG nova.compute.manager [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.989978] env[61006]: DEBUG nova.network.neutron [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.092229] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337004, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.228674] env[61006]: INFO nova.compute.manager [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Took 22.61 seconds to build instance. [ 834.279712] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ce1547-4ee0-15ab-75b9-45361eed0dc6, 'name': SearchDatastore_Task, 'duration_secs': 0.065143} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.280467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.280796] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.281230] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.281486] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.281717] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.282107] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-865194be-7b41-4e95-a855-1f364346ae78 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.292888] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.293087] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.293809] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c021af1-ba64-4607-8d57-f6282fa3131b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.301822] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 834.301822] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d89b76-849b-831c-aa65-5eab7e922262" [ 834.301822] env[61006]: _type = "Task" [ 834.301822] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.309350] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d89b76-849b-831c-aa65-5eab7e922262, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.324438] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478edb94-b13d-4c75-804d-95fda50e3a2e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.331107] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95abfaea-d910-4e7e-813d-9c3f1345b87b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.364700] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9334a7d-e7f2-4457-8c70-c48def9241c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.371792] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc4f396-0175-4f9d-ab54-bc1ae848d1a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.385545] env[61006]: DEBUG nova.compute.provider_tree [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.476329] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updated VIF entry in instance network info cache for port e8f20725-aaab-4cc2-87d5-89688205c617. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.476732] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [{"id": "e8f20725-aaab-4cc2-87d5-89688205c617", "address": "fa:16:3e:8b:b1:62", "network": {"id": "9a269f3a-edae-4502-9c28-a778a68a33b3", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1040168596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "064b30c918ac499e9838488a08256d29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "44ed8f45-cb8e-40e7-ac70-a7f386a7d2c2", "external-id": "nsx-vlan-transportzone-268", "segmentation_id": 268, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8f20725-aa", "ovs_interfaceid": "e8f20725-aaab-4cc2-87d5-89688205c617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.480161] env[61006]: DEBUG nova.network.neutron [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updated VIF entry in instance network info cache for port 3d9ba3b1-a27e-4002-bd16-cfd881fca452. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.480704] env[61006]: DEBUG nova.network.neutron [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.583793] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521547} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.584092] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 72c3e37b-2eac-41d4-8308-0a6466c2dd24/72c3e37b-2eac-41d4-8308-0a6466c2dd24.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.584314] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.584602] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87fc03ec-8c4c-429f-849a-0e2ec17f0635 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.591537] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 834.591537] env[61006]: value = "task-1337005" [ 834.591537] env[61006]: _type = "Task" [ 834.591537] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.600455] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337005, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.731216] env[61006]: DEBUG oslo_concurrency.lockutils [None req-99c911ad-f7ef-4667-92e9-f5a1d0d2c661 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.990s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.813611] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d89b76-849b-831c-aa65-5eab7e922262, 'name': SearchDatastore_Task, 'duration_secs': 0.008327} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.814466] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-277fc569-50d2-4bf0-9216-45d952e654cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.819724] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 834.819724] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524147c1-4427-e378-6706-4c9aba85a63d" [ 834.819724] env[61006]: _type = "Task" [ 834.819724] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.827840] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524147c1-4427-e378-6706-4c9aba85a63d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.922214] env[61006]: DEBUG nova.scheduler.client.report [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 834.922571] env[61006]: DEBUG nova.compute.provider_tree [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 86 to 87 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 834.922678] env[61006]: DEBUG nova.compute.provider_tree [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.977572] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.980020] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Releasing lock "refresh_cache-88e2bdc9-ab73-4e23-94b5-a45046835144" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.980435] env[61006]: DEBUG nova.compute.manager [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Received event network-changed-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.980435] env[61006]: DEBUG nova.compute.manager [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Refreshing instance network info cache due to event network-changed-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.980658] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Acquiring lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.980806] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Acquired lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.980972] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Refreshing network info cache for port 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.982978] env[61006]: DEBUG oslo_concurrency.lockutils [req-49cdb718-618e-499f-9d7d-97de9a6336e4 req-81ae2c27-19d7-462f-80f9-ed7527ee7378 service nova] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.003025] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.003513] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.003552] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.003814] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.004013] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.004206] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.004529] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.004783] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.004935] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.005173] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.005400] env[61006]: DEBUG nova.virt.hardware [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.005679] env[61006]: DEBUG nova.network.neutron [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.007668] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243a29f7-58be-43ab-94ac-4672af129179 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.016716] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca84047-24c5-4233-9038-cac13dacd4c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.102523] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337005, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.240622} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.102817] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.103608] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7377b93b-3f35-4492-8c1a-1a7ab005fad0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.125910] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 72c3e37b-2eac-41d4-8308-0a6466c2dd24/72c3e37b-2eac-41d4-8308-0a6466c2dd24.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.126221] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0d984bc-f208-47a4-9003-c5b7af444990 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.148394] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 835.148394] env[61006]: value = "task-1337006" [ 835.148394] env[61006]: _type = "Task" [ 835.148394] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.155810] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337006, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.233941] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.246722] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "598c0d72-d679-49a8-b17c-f5f341c205e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.247047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.247298] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.247540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.247757] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.249886] env[61006]: INFO nova.compute.manager [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Terminating instance [ 835.252030] env[61006]: DEBUG nova.compute.manager [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.252030] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.252821] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6495803-adf0-472c-b291-d220c9b70190 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.260792] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.260894] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12e8ba95-b0ea-4329-b8ed-9aa12c5acb34 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.266595] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 835.266595] env[61006]: value = "task-1337007" [ 835.266595] env[61006]: _type = "Task" [ 835.266595] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.276049] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.330786] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524147c1-4427-e378-6706-4c9aba85a63d, 'name': SearchDatastore_Task, 'duration_secs': 0.061555} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.331164] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.331545] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.331898] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e2cf05b-9210-4c85-a1f9-9c84b06750cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.338939] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 835.338939] env[61006]: value = "task-1337008" [ 835.338939] env[61006]: _type = "Task" [ 835.338939] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.346801] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.355490] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "29c41817-2189-4622-8f35-86f61eb34bed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.355825] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.356153] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "29c41817-2189-4622-8f35-86f61eb34bed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.356361] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.356570] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.359258] env[61006]: INFO nova.compute.manager [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Terminating instance [ 835.361215] env[61006]: DEBUG nova.compute.manager [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 835.361451] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.362293] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2a6837-712c-43f5-ac22-74d9011550e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.370050] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.370427] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55b2943e-1428-4a02-9081-2116d8c72240 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.376860] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 835.376860] env[61006]: value = "task-1337009" [ 835.376860] env[61006]: _type = "Task" [ 835.376860] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.388923] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.412986] env[61006]: DEBUG nova.compute.manager [req-b64371a0-f434-4781-a951-20ac74348364 req-34a018c7-d1bf-45c5-b481-48bff78df5b4 service nova] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Received event network-vif-deleted-e8f20725-aaab-4cc2-87d5-89688205c617 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.430038] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.430038] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.433386] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.252s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.436278] env[61006]: INFO nova.compute.claims [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.449753] env[61006]: DEBUG nova.compute.manager [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Received event network-vif-plugged-b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.449753] env[61006]: DEBUG oslo_concurrency.lockutils [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] Acquiring lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.449753] env[61006]: DEBUG oslo_concurrency.lockutils [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.449753] env[61006]: DEBUG oslo_concurrency.lockutils [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.449753] env[61006]: DEBUG nova.compute.manager [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] No waiting events found dispatching network-vif-plugged-b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.450068] env[61006]: WARNING nova.compute.manager [req-9fe46426-5478-4327-8a06-f5e7aa2b5db7 req-abfa4e0b-218d-4dec-b9f0-977bf2dce8b2 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Received unexpected event network-vif-plugged-b6ca963a-897a-4678-b1ed-75c5a1099ee2 for instance with vm_state building and task_state spawning. [ 835.450522] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Successfully updated port: b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.512382] env[61006]: INFO nova.compute.manager [-] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Took 1.52 seconds to deallocate network for instance. [ 835.660796] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337006, 'name': ReconfigVM_Task, 'duration_secs': 0.255261} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.661139] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 72c3e37b-2eac-41d4-8308-0a6466c2dd24/72c3e37b-2eac-41d4-8308-0a6466c2dd24.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.661781] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-614e4482-063b-45e5-bb1e-b9883eafc5ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.669168] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 835.669168] env[61006]: value = "task-1337010" [ 835.669168] env[61006]: _type = "Task" [ 835.669168] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.681984] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337010, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.763990] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.778017] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337007, 'name': PowerOffVM_Task, 'duration_secs': 0.344689} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.778346] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.778525] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.779278] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59db9149-57ba-4280-ba52-fc2ee7d75deb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.836027] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.836280] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.836463] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleting the datastore file [datastore2] 598c0d72-d679-49a8-b17c-f5f341c205e8 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.836748] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc4205e7-0804-4af4-b8cc-62a1847bc84a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.846385] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 835.846385] env[61006]: value = "task-1337012" [ 835.846385] env[61006]: _type = "Task" [ 835.846385] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.847087] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updated VIF entry in instance network info cache for port 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.847432] env[61006]: DEBUG nova.network.neutron [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updating instance_info_cache with network_info: [{"id": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "address": "fa:16:3e:f0:17:57", "network": {"id": "eba91a43-0fae-4db5-99f1-134ac43cea76", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1804121233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24ba46b1b91f46479841ea9db9dce16f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7043ca7a-807c-4c7b-b646-23ffece188b2", "external-id": "nsx-vlan-transportzone-619", "segmentation_id": 619, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0652e5ba-2f", "ovs_interfaceid": "0652e5ba-2f85-49a7-85aa-2f0bc1757b4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.854971] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483129} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.855687] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.855877] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.856164] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15eda6ea-5238-4f86-beeb-a402b2899b9e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.860940] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337012, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.865070] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 835.865070] env[61006]: value = "task-1337013" [ 835.865070] env[61006]: _type = "Task" [ 835.865070] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.872335] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337013, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.887781] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337009, 'name': PowerOffVM_Task, 'duration_secs': 0.240387} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.888531] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.888756] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.889292] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b70e3f40-ae4e-418d-a29b-038cbd4a94fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.935991] env[61006]: DEBUG nova.compute.utils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.937622] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.937836] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 835.944243] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.944461] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.944647] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleting the datastore file [datastore2] 29c41817-2189-4622-8f35-86f61eb34bed {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.945113] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad7711b0-bc54-4aaf-b20c-fab296c3aada {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.951870] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 835.951870] env[61006]: value = "task-1337015" [ 835.951870] env[61006]: _type = "Task" [ 835.951870] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.955768] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.955900] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquired lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.956054] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 835.962398] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.998599] env[61006]: DEBUG nova.policy [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc2220b969c942efba38ebd09ec63ed6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbbb1cf0f7974ccdb451078aaa448272', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.020515] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.180312] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337010, 'name': Rename_Task, 'duration_secs': 0.218898} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.180673] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.181148] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1669c268-2091-4d9e-9ef8-f5009a71b499 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.187730] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 836.187730] env[61006]: value = "task-1337016" [ 836.187730] env[61006]: _type = "Task" [ 836.187730] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.196438] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.353025] env[61006]: DEBUG oslo_concurrency.lockutils [req-a34bbdea-41ae-431b-b835-4f9c84b71c85 req-973245fe-a3ac-42fc-9be3-0fd5dd7e7fba service nova] Releasing lock "refresh_cache-42b92d52-d1f0-48ff-94b6-6164b387456c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.359199] env[61006]: DEBUG oslo_vmware.api [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337012, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145382} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.359610] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.359910] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.360216] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.360509] env[61006]: INFO nova.compute.manager [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 836.360882] env[61006]: DEBUG oslo.service.loopingcall [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.361196] env[61006]: DEBUG nova.compute.manager [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.361407] env[61006]: DEBUG nova.network.neutron [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.376238] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337013, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072926} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.376550] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.377684] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602236ee-a955-4523-9f65-9e0994355cce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.399978] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.400355] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1278b18a-8d75-407b-a2de-bd5910948adf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.420160] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 836.420160] env[61006]: value = "task-1337017" [ 836.420160] env[61006]: _type = "Task" [ 836.420160] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.428012] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.441962] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.463032] env[61006]: DEBUG oslo_vmware.api [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137529} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.463593] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.463802] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.464429] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.464429] env[61006]: INFO nova.compute.manager [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Took 1.10 seconds to destroy the instance on the hypervisor. [ 836.464504] env[61006]: DEBUG oslo.service.loopingcall [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.464691] env[61006]: DEBUG nova.compute.manager [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.464789] env[61006]: DEBUG nova.network.neutron [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.506350] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Successfully created port: ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.546660] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.697998] env[61006]: DEBUG oslo_vmware.api [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337016, 'name': PowerOnVM_Task, 'duration_secs': 0.459017} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.700858] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.701113] env[61006]: INFO nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Took 8.85 seconds to spawn the instance on the hypervisor. [ 836.701272] env[61006]: DEBUG nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.702558] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d72e88-9a88-414c-b4e6-ec33f6c20dfc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.791522] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f500682-8a8d-4703-8710-96b873abc348 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.799433] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f824b31-ca3c-461f-a359-13b3db6b7b67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.834462] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1bd35a-f414-4b8b-acc2-8f400321702d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.842430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24772f90-ce3a-4d77-9175-943670621811 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.857385] env[61006]: DEBUG nova.compute.provider_tree [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.925557] env[61006]: DEBUG nova.network.neutron [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Updating instance_info_cache with network_info: [{"id": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "address": "fa:16:3e:75:7e:71", "network": {"id": "93d47403-022b-4d01-9499-15990e766b46", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1509926307-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0df410358ec405088642ab7614b2293", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ca963a-89", "ovs_interfaceid": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.934441] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337017, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.224722] env[61006]: INFO nova.compute.manager [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Took 24.78 seconds to build instance. [ 837.360329] env[61006]: DEBUG nova.scheduler.client.report [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.384741] env[61006]: DEBUG nova.network.neutron [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.433019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Releasing lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.433019] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Instance network_info: |[{"id": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "address": "fa:16:3e:75:7e:71", "network": {"id": "93d47403-022b-4d01-9499-15990e766b46", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1509926307-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0df410358ec405088642ab7614b2293", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ca963a-89", "ovs_interfaceid": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.433352] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337017, 'name': ReconfigVM_Task, 'duration_secs': 0.55379} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.433352] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:7e:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a31c4b8-5b72-4f32-aab3-c4e963e684dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6ca963a-897a-4678-b1ed-75c5a1099ee2', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.439550] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Creating folder: Project (f0df410358ec405088642ab7614b2293). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.439829] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.440444] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b4542d5-9e0f-4a95-b9e5-5f488970d687 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.442149] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f27d4765-d688-4a29-af9d-64208c9ac9ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.446905] env[61006]: DEBUG nova.compute.manager [req-7382c2ae-7a26-4ff1-bbfe-9b7f7f1f2b15 req-47426b3b-f76d-4067-a2a5-5064e8cc9fee service nova] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Received event network-vif-deleted-b2c4d5fa-8972-4475-af28-caa8b5ffce39 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.447107] env[61006]: DEBUG nova.compute.manager [req-7382c2ae-7a26-4ff1-bbfe-9b7f7f1f2b15 req-47426b3b-f76d-4067-a2a5-5064e8cc9fee service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Received event network-vif-deleted-608225ce-37a4-4c41-ba85-f2e0575b97d1 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.447326] env[61006]: INFO nova.compute.manager [req-7382c2ae-7a26-4ff1-bbfe-9b7f7f1f2b15 req-47426b3b-f76d-4067-a2a5-5064e8cc9fee service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Neutron deleted interface 608225ce-37a4-4c41-ba85-f2e0575b97d1; detaching it from the instance and deleting it from the info cache [ 837.447438] env[61006]: DEBUG nova.network.neutron [req-7382c2ae-7a26-4ff1-bbfe-9b7f7f1f2b15 req-47426b3b-f76d-4067-a2a5-5064e8cc9fee service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.450215] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 837.450215] env[61006]: value = "task-1337018" [ 837.450215] env[61006]: _type = "Task" [ 837.450215] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.451146] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.457122] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Created folder: Project (f0df410358ec405088642ab7614b2293) in parent group-v285275. [ 837.457320] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Creating folder: Instances. Parent ref: group-v285330. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.457548] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dded11c8-5b93-4c3f-b933-0c4708a60d70 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.465195] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337018, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.468255] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Created folder: Instances in parent group-v285330. [ 837.468479] env[61006]: DEBUG oslo.service.loopingcall [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.470232] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.470985] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a42b7d76-7335-467e-bf35-c2a230e67b63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.488013] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.488266] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.488434] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.488621] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.488771] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.488922] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.489157] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.489310] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.489584] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.489637] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.489815] env[61006]: DEBUG nova.virt.hardware [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.491249] env[61006]: DEBUG nova.network.neutron [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.493565] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efab8469-a029-4827-b8c4-ee82bbfabca8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.496812] env[61006]: DEBUG nova.compute.manager [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Received event network-changed-b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.496999] env[61006]: DEBUG nova.compute.manager [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Refreshing instance network info cache due to event network-changed-b6ca963a-897a-4678-b1ed-75c5a1099ee2. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.497233] env[61006]: DEBUG oslo_concurrency.lockutils [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] Acquiring lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.497402] env[61006]: DEBUG oslo_concurrency.lockutils [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] Acquired lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.497579] env[61006]: DEBUG nova.network.neutron [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Refreshing network info cache for port b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.507931] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793f8074-a0b0-4a5c-a54e-0a2b8774693a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.512227] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.512227] env[61006]: value = "task-1337021" [ 837.512227] env[61006]: _type = "Task" [ 837.512227] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.528679] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337021, 'name': CreateVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.727954] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15b35444-dffd-44f2-aeac-eeacbed774b3 tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.881s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.866932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.867542] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.870213] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.388s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.872017] env[61006]: INFO nova.compute.claims [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.888399] env[61006]: INFO nova.compute.manager [-] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Took 1.53 seconds to deallocate network for instance. [ 837.950236] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-140a6975-4fe9-4c6f-b515-1dc963051290 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.962960] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337018, 'name': Rename_Task, 'duration_secs': 0.148568} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.964197] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.964482] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5706ca9f-c808-4589-9d6b-940b27f1b7ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.968863] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bda4ce-72c5-46e8-b502-4a3e8cc48656 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.988280] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 837.988280] env[61006]: value = "task-1337022" [ 837.988280] env[61006]: _type = "Task" [ 837.988280] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.993317] env[61006]: INFO nova.compute.manager [-] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Took 1.53 seconds to deallocate network for instance. [ 838.006627] env[61006]: DEBUG nova.compute.manager [req-7382c2ae-7a26-4ff1-bbfe-9b7f7f1f2b15 req-47426b3b-f76d-4067-a2a5-5064e8cc9fee service nova] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Detach interface failed, port_id=608225ce-37a4-4c41-ba85-f2e0575b97d1, reason: Instance 29c41817-2189-4622-8f35-86f61eb34bed could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 838.010130] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337022, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.021667] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337021, 'name': CreateVM_Task, 'duration_secs': 0.355782} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.021667] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.021794] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.021953] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.023157] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.023157] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a59fc7d2-3e74-40d2-9580-cb97474dd4e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.027152] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 838.027152] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fc340b-fe7f-ed98-e8c3-ccee866e3394" [ 838.027152] env[61006]: _type = "Task" [ 838.027152] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.036260] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fc340b-fe7f-ed98-e8c3-ccee866e3394, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.104701] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Successfully updated port: ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.230832] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.237049] env[61006]: DEBUG nova.network.neutron [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Updated VIF entry in instance network info cache for port b6ca963a-897a-4678-b1ed-75c5a1099ee2. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.237427] env[61006]: DEBUG nova.network.neutron [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Updating instance_info_cache with network_info: [{"id": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "address": "fa:16:3e:75:7e:71", "network": {"id": "93d47403-022b-4d01-9499-15990e766b46", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1509926307-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0df410358ec405088642ab7614b2293", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a31c4b8-5b72-4f32-aab3-c4e963e684dd", "external-id": "nsx-vlan-transportzone-805", "segmentation_id": 805, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6ca963a-89", "ovs_interfaceid": "b6ca963a-897a-4678-b1ed-75c5a1099ee2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.376940] env[61006]: DEBUG nova.compute.utils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.380530] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.380713] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.395956] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.425281] env[61006]: DEBUG nova.policy [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc2220b969c942efba38ebd09ec63ed6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbbb1cf0f7974ccdb451078aaa448272', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.504255] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337022, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.509561] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.539181] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fc340b-fe7f-ed98-e8c3-ccee866e3394, 'name': SearchDatastore_Task, 'duration_secs': 0.00994} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.539503] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.539744] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.540151] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.540333] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.540541] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.540819] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49551c73-82f6-455c-9d7a-980449ad0f13 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.550540] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.551794] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.551794] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41d6d990-5234-464d-b845-356f239657a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.557511] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 838.557511] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528ee50a-6e8f-6acb-6266-45618ef37719" [ 838.557511] env[61006]: _type = "Task" [ 838.557511] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.567412] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528ee50a-6e8f-6acb-6266-45618ef37719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.607616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.607616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.607616] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.733837] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Successfully created port: 63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.739553] env[61006]: DEBUG oslo_concurrency.lockutils [req-7ba2f403-d630-49c4-aa62-7d2e1fa2ba30 req-a2155042-873e-476c-86ae-d5da880ac626 service nova] Releasing lock "refresh_cache-28b83ea7-5433-472d-9e47-f73a4f2fb389" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.758199] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.881227] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.999419] env[61006]: DEBUG oslo_vmware.api [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337022, 'name': PowerOnVM_Task, 'duration_secs': 0.653845} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.999660] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.999919] env[61006]: INFO nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 8.60 seconds to spawn the instance on the hypervisor. [ 839.000143] env[61006]: DEBUG nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.001087] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b35c33b-6353-4e58-851e-1831168d8938 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.069142] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528ee50a-6e8f-6acb-6266-45618ef37719, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.072422] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75b36103-7bf0-42ba-bed3-696e38567d90 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.076762] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 839.076762] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52006016-4b0a-f229-92f2-389e34db47dd" [ 839.076762] env[61006]: _type = "Task" [ 839.076762] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.085702] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52006016-4b0a-f229-92f2-389e34db47dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.158929] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.191937] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a638e2e5-9004-485b-a480-51dc7f25e1eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.199818] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f33def1-c30c-435d-be1a-11ea03cc8450 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.232737] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d354be3b-d3fc-4160-bd02-76e82fd37503 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.243118] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd63506f-4688-4382-8dfb-84e34f45441b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.257808] env[61006]: DEBUG nova.compute.provider_tree [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.325534] env[61006]: DEBUG nova.network.neutron [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [{"id": "ba96f533-658c-49f4-a9a4-48e129d5c197", "address": "fa:16:3e:bb:35:70", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba96f533-65", "ovs_interfaceid": "ba96f533-658c-49f4-a9a4-48e129d5c197", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.524420] env[61006]: INFO nova.compute.manager [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 26.05 seconds to build instance. [ 839.540146] env[61006]: DEBUG nova.compute.manager [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Received event network-vif-plugged-ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.540379] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Acquiring lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.540591] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.540798] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.541047] env[61006]: DEBUG nova.compute.manager [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] No waiting events found dispatching network-vif-plugged-ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.541614] env[61006]: WARNING nova.compute.manager [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Received unexpected event network-vif-plugged-ba96f533-658c-49f4-a9a4-48e129d5c197 for instance with vm_state building and task_state spawning. [ 839.541614] env[61006]: DEBUG nova.compute.manager [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Received event network-changed-ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.541614] env[61006]: DEBUG nova.compute.manager [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Refreshing instance network info cache due to event network-changed-ba96f533-658c-49f4-a9a4-48e129d5c197. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.541755] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Acquiring lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.586538] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52006016-4b0a-f229-92f2-389e34db47dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010229} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.586811] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.587084] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 28b83ea7-5433-472d-9e47-f73a4f2fb389/28b83ea7-5433-472d-9e47-f73a4f2fb389.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.587391] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc45b062-7f1a-4213-9d1a-b5368b8dc33b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.594122] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 839.594122] env[61006]: value = "task-1337023" [ 839.594122] env[61006]: _type = "Task" [ 839.594122] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.602434] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.762626] env[61006]: DEBUG nova.scheduler.client.report [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.828648] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.829026] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Instance network_info: |[{"id": "ba96f533-658c-49f4-a9a4-48e129d5c197", "address": "fa:16:3e:bb:35:70", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba96f533-65", "ovs_interfaceid": "ba96f533-658c-49f4-a9a4-48e129d5c197", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 839.829385] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Acquired lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.829576] env[61006]: DEBUG nova.network.neutron [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Refreshing network info cache for port ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.830921] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:35:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba96f533-658c-49f4-a9a4-48e129d5c197', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.838758] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating folder: Project (cbbb1cf0f7974ccdb451078aaa448272). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.839892] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a3d63f5-82e2-4337-92b6-3ffde110601e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.852034] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created folder: Project (cbbb1cf0f7974ccdb451078aaa448272) in parent group-v285275. [ 839.852675] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating folder: Instances. Parent ref: group-v285333. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.852675] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc064f81-505a-4ef0-882f-ab8f31a70889 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.862143] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created folder: Instances in parent group-v285333. [ 839.862451] env[61006]: DEBUG oslo.service.loopingcall [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.862690] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 839.862918] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1e2b31c-d259-4f6f-bef6-bc506fc380a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.884118] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.884118] env[61006]: value = "task-1337026" [ 839.884118] env[61006]: _type = "Task" [ 839.884118] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.892822] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337026, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.896071] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.934995] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.935683] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.935895] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.936131] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.936304] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.936505] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.936757] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.936996] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.937206] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.937406] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.937596] env[61006]: DEBUG nova.virt.hardware [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.938614] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1699ba0d-5439-4dae-8624-50ac51057332 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.948400] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f574bdb-d4e0-45e9-b678-67e92788aadc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.027409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7c9e5c95-d109-4814-83c4-eb1b33bced16 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.131s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.106802] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337023, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.150161] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.150653] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.150916] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.151167] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.151537] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.153981] env[61006]: INFO nova.compute.manager [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Terminating instance [ 840.156279] env[61006]: DEBUG nova.compute.manager [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.156511] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.157457] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2c5d84-e02b-4995-8084-89f8888b7f45 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.168974] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.169322] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf81716b-ed75-45e7-8b44-6cab18c30ace {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.182493] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 840.182493] env[61006]: value = "task-1337027" [ 840.182493] env[61006]: _type = "Task" [ 840.182493] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.189520] env[61006]: DEBUG nova.compute.manager [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Received event network-changed-3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.189802] env[61006]: DEBUG nova.compute.manager [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Refreshing instance network info cache due to event network-changed-3d9ba3b1-a27e-4002-bd16-cfd881fca452. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.190046] env[61006]: DEBUG oslo_concurrency.lockutils [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.190236] env[61006]: DEBUG oslo_concurrency.lockutils [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.190413] env[61006]: DEBUG nova.network.neutron [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Refreshing network info cache for port 3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.201601] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.268600] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.269151] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.275019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.319s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.275019] env[61006]: INFO nova.compute.claims [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.400184] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337026, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.400184] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Successfully updated port: 63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.530746] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.608990] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669466} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.608990] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 28b83ea7-5433-472d-9e47-f73a4f2fb389/28b83ea7-5433-472d-9e47-f73a4f2fb389.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.608990] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.608990] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e587d643-b464-4882-b42c-4cc1ac053b30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.614893] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 840.614893] env[61006]: value = "task-1337028" [ 840.614893] env[61006]: _type = "Task" [ 840.614893] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.619144] env[61006]: DEBUG nova.network.neutron [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updated VIF entry in instance network info cache for port ba96f533-658c-49f4-a9a4-48e129d5c197. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.619700] env[61006]: DEBUG nova.network.neutron [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [{"id": "ba96f533-658c-49f4-a9a4-48e129d5c197", "address": "fa:16:3e:bb:35:70", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba96f533-65", "ovs_interfaceid": "ba96f533-658c-49f4-a9a4-48e129d5c197", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.625222] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.692547] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337027, 'name': PowerOffVM_Task, 'duration_secs': 0.433534} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.692547] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.692547] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.692744] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bfd0770-982a-4b9d-a3c5-bc8dda4a8383 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.775093] env[61006]: DEBUG nova.compute.utils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.776479] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.776651] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.794467] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.794688] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.794867] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Deleting the datastore file [datastore2] 72c3e37b-2eac-41d4-8308-0a6466c2dd24 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.795146] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f715f40-81af-4014-8db3-35bf4e370dac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.802201] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for the task: (returnval){ [ 840.802201] env[61006]: value = "task-1337030" [ 840.802201] env[61006]: _type = "Task" [ 840.802201] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.809819] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.822562] env[61006]: DEBUG nova.policy [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d13e18f459843429bbc204da327480f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4693570ceab84d1babfa6f8f24447fcd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.898316] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337026, 'name': CreateVM_Task, 'duration_secs': 0.606682} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.898524] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.899153] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.899367] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.899693] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.899958] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c40c67a8-2ae9-4aa5-b356-dcd85a2c6265 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.906885] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.906885] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.906885] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.906885] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 840.906885] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5206161f-992b-ee76-86a6-f6d6f27372f1" [ 840.906885] env[61006]: _type = "Task" [ 840.906885] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.914517] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5206161f-992b-ee76-86a6-f6d6f27372f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.953876] env[61006]: DEBUG nova.network.neutron [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updated VIF entry in instance network info cache for port 3d9ba3b1-a27e-4002-bd16-cfd881fca452. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 840.954282] env[61006]: DEBUG nova.network.neutron [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.054765] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.075436] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Successfully created port: 6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.125256] env[61006]: DEBUG oslo_concurrency.lockutils [req-dd2f2745-9efe-4139-963c-b4bb800eb0a5 req-8dd0a96a-cde6-4d14-9ccb-8369f0a432ca service nova] Releasing lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.125688] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067218} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.125932] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.126736] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dde8c13-7272-49b2-8d24-9ebfa34cd486 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.150251] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 28b83ea7-5433-472d-9e47-f73a4f2fb389/28b83ea7-5433-472d-9e47-f73a4f2fb389.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.150584] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eca1f5b2-4104-4405-97fa-a8605dfa5e83 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.170703] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 841.170703] env[61006]: value = "task-1337031" [ 841.170703] env[61006]: _type = "Task" [ 841.170703] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.179052] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.207600] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.207600] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.282043] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.313503] env[61006]: DEBUG oslo_vmware.api [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Task: {'id': task-1337030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161791} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.313785] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.314314] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.314558] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.314749] env[61006]: INFO nova.compute.manager [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Took 1.16 seconds to destroy the instance on the hypervisor. [ 841.315018] env[61006]: DEBUG oslo.service.loopingcall [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.318069] env[61006]: DEBUG nova.compute.manager [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.322016] env[61006]: DEBUG nova.network.neutron [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.422925] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5206161f-992b-ee76-86a6-f6d6f27372f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010919} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.423436] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.424144] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.424265] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.424459] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.424773] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.426058] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Successfully created port: cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.429316] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b1dde4d-02d5-4caa-9990-7fcf5d491161 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.447118] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.447337] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.448130] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7585ea0d-1502-4eea-a4c1-a130e63f1cb7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.461281] env[61006]: DEBUG oslo_concurrency.lockutils [req-5482cd63-0510-4171-b12a-2165cc51e57a req-31369bd2-c4fc-4c30-8aa4-3e056d53c01b service nova] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.461750] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 841.461750] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52395f21-6a19-375c-b3fb-d9b38935ef49" [ 841.461750] env[61006]: _type = "Task" [ 841.461750] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.465551] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.472373] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52395f21-6a19-375c-b3fb-d9b38935ef49, 'name': SearchDatastore_Task, 'duration_secs': 0.008299} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.476979] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb940b19-606d-4b4e-9cf3-97c847655db2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.482950] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 841.482950] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5260253d-e0d4-f2d2-0bd8-dd16bdf60639" [ 841.482950] env[61006]: _type = "Task" [ 841.482950] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.494037] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5260253d-e0d4-f2d2-0bd8-dd16bdf60639, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.670144] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b756ef7-15d1-4426-9cfe-6761285dd675 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.683737] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6514df16-94c4-4a32-9aa9-9d07994092df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.687153] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337031, 'name': ReconfigVM_Task, 'duration_secs': 0.279519} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.690235] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 28b83ea7-5433-472d-9e47-f73a4f2fb389/28b83ea7-5433-472d-9e47-f73a4f2fb389.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.691270] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22460470-2f0e-4292-a7ed-46f2244147c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.726197] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4aebe3-d6e8-4844-965f-ced6dfbd0a29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.728377] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.728525] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 841.728650] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 841.729993] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 841.729993] env[61006]: value = "task-1337032" [ 841.729993] env[61006]: _type = "Task" [ 841.729993] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.738016] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7fa2ba-35a1-4542-8a3c-2b5db56e3afa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.746023] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337032, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.758385] env[61006]: DEBUG nova.compute.provider_tree [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.781650] env[61006]: DEBUG nova.compute.manager [req-f6565560-505d-44de-a048-800f601e5526 req-f7157bd0-9553-4ea0-8e6f-8d6e9001e6d9 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Received event network-vif-deleted-52eb94ea-4588-467f-9b0c-273dbbae4f8b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.781891] env[61006]: INFO nova.compute.manager [req-f6565560-505d-44de-a048-800f601e5526 req-f7157bd0-9553-4ea0-8e6f-8d6e9001e6d9 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Neutron deleted interface 52eb94ea-4588-467f-9b0c-273dbbae4f8b; detaching it from the instance and deleting it from the info cache [ 841.781999] env[61006]: DEBUG nova.network.neutron [req-f6565560-505d-44de-a048-800f601e5526 req-f7157bd0-9553-4ea0-8e6f-8d6e9001e6d9 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.932321] env[61006]: DEBUG nova.network.neutron [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Updating instance_info_cache with network_info: [{"id": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "address": "fa:16:3e:3f:75:e4", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63ee7dc4-4d", "ovs_interfaceid": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.993596] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5260253d-e0d4-f2d2-0bd8-dd16bdf60639, 'name': SearchDatastore_Task, 'duration_secs': 0.00885} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.994173] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.994173] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/0111452e-1b4f-499c-932d-f31364d1a14c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 841.994437] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4176ba18-3796-4863-ab54-219be35e4f0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.000606] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 842.000606] env[61006]: value = "task-1337033" [ 842.000606] env[61006]: _type = "Task" [ 842.000606] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.008054] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337033, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.133140] env[61006]: DEBUG nova.network.neutron [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.235795] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Skipping network cache update for instance because it is being deleted. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 842.235988] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 842.236135] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 842.236263] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 842.236388] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 842.236513] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 842.240054] env[61006]: DEBUG nova.compute.manager [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Received event network-vif-plugged-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.240302] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Acquiring lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.240513] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.240688] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.240857] env[61006]: DEBUG nova.compute.manager [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] No waiting events found dispatching network-vif-plugged-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.241055] env[61006]: WARNING nova.compute.manager [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Received unexpected event network-vif-plugged-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb for instance with vm_state building and task_state spawning. [ 842.241266] env[61006]: DEBUG nova.compute.manager [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Received event network-changed-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.241451] env[61006]: DEBUG nova.compute.manager [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Refreshing instance network info cache due to event network-changed-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.241633] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Acquiring lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.253305] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337032, 'name': Rename_Task, 'duration_secs': 0.135786} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.253631] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.253958] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b60e58d-e734-4218-987a-816913c1e0df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.263550] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 842.263550] env[61006]: value = "task-1337034" [ 842.263550] env[61006]: _type = "Task" [ 842.263550] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.273028] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337034, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.280516] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.280713] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.280836] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Forcefully refreshing network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 842.280983] env[61006]: DEBUG nova.objects.instance [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lazy-loading 'info_cache' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.284563] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5611a0e1-5cf8-4334-92b1-80d5d543916c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.295200] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e11838f-20c2-4df7-bdb1-e03eaa4aaf22 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.307200] env[61006]: DEBUG nova.scheduler.client.report [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 842.307528] env[61006]: DEBUG nova.compute.provider_tree [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 87 to 88 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 842.307705] env[61006]: DEBUG nova.compute.provider_tree [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.312864] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.332947] env[61006]: DEBUG nova.compute.manager [req-f6565560-505d-44de-a048-800f601e5526 req-f7157bd0-9553-4ea0-8e6f-8d6e9001e6d9 service nova] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Detach interface failed, port_id=52eb94ea-4588-467f-9b0c-273dbbae4f8b, reason: Instance 72c3e37b-2eac-41d4-8308-0a6466c2dd24 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 842.343100] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.343100] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.343100] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.343336] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.343368] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.344034] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.344034] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.344034] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.344187] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.344335] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.344513] env[61006]: DEBUG nova.virt.hardware [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.345709] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25762bcd-e482-4b98-ab8f-c8f0805d0dbc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.355108] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174c7d1a-8883-469c-9be3-082461dac433 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.435280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.436027] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Instance network_info: |[{"id": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "address": "fa:16:3e:3f:75:e4", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63ee7dc4-4d", "ovs_interfaceid": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.436869] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Acquired lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.436869] env[61006]: DEBUG nova.network.neutron [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Refreshing network info cache for port 63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.438566] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:75:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.447711] env[61006]: DEBUG oslo.service.loopingcall [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.448288] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.448544] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6eec7b44-7b05-4a26-9257-cae7b6455090 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.472392] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.472392] env[61006]: value = "task-1337035" [ 842.472392] env[61006]: _type = "Task" [ 842.472392] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.483195] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337035, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.510924] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337033, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.636068] env[61006]: INFO nova.compute.manager [-] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Took 1.32 seconds to deallocate network for instance. [ 842.775196] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337034, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.817164] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.817834] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.821439] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.455s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.822070] env[61006]: DEBUG nova.objects.instance [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lazy-loading 'resources' on Instance uuid ea7d5d49-ac76-4f2e-9456-912cf466fcc2 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.983916] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337035, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.010927] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337033, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526322} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.010927] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/0111452e-1b4f-499c-932d-f31364d1a14c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.011127] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.011738] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b0796ab-6277-4699-87da-11e7ef0a3267 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.017824] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 843.017824] env[61006]: value = "task-1337036" [ 843.017824] env[61006]: _type = "Task" [ 843.017824] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.024888] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.142928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.275459] env[61006]: DEBUG oslo_vmware.api [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337034, 'name': PowerOnVM_Task, 'duration_secs': 0.823467} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.275736] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.275945] env[61006]: INFO nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Took 8.30 seconds to spawn the instance on the hypervisor. [ 843.276146] env[61006]: DEBUG nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.276933] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a65ccc4-6caf-4e0d-bd98-7f0650c86dce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.280190] env[61006]: DEBUG nova.network.neutron [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Updated VIF entry in instance network info cache for port 63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.280523] env[61006]: DEBUG nova.network.neutron [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Updating instance_info_cache with network_info: [{"id": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "address": "fa:16:3e:3f:75:e4", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63ee7dc4-4d", "ovs_interfaceid": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.326549] env[61006]: DEBUG nova.compute.utils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.327882] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.332283] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.384819] env[61006]: DEBUG nova.policy [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.484850] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337035, 'name': CreateVM_Task, 'duration_secs': 0.560729} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.487384] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.488415] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.488663] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.489116] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.489681] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c33994d-7eb5-4956-a9fa-5602d427340e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.494746] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 843.494746] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d3676-a15d-2057-10f7-af7cf91ca913" [ 843.494746] env[61006]: _type = "Task" [ 843.494746] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.506782] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d3676-a15d-2057-10f7-af7cf91ca913, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.525853] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Successfully updated port: 6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.531921] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073753} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.532599] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.536182] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818c3af8-8369-4737-b5e8-3693f66a30f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.561087] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/0111452e-1b4f-499c-932d-f31364d1a14c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.564872] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dad40822-5be3-4e1d-a8de-7f920198ed53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.585756] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 843.585756] env[61006]: value = "task-1337037" [ 843.585756] env[61006]: _type = "Task" [ 843.585756] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.598259] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337037, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.683016] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b72e4c1-8c1a-4e8d-b7d1-7e14672a3ddc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.690160] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cb62cc-7c4e-4c80-be72-c2f9530bc4bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.720166] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f5c6cf-3b80-415a-9e13-8ed6073de544 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.728009] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cafcbc0-4d73-4f9d-af34-96d479aae193 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.744084] env[61006]: DEBUG nova.compute.provider_tree [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.769325] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Successfully created port: b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.786158] env[61006]: DEBUG oslo_concurrency.lockutils [req-4bfbd13f-7efd-4e53-84a9-35e18651a8c4 req-449711d5-6036-44d8-b6a0-f0e7eebd6117 service nova] Releasing lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.806485] env[61006]: INFO nova.compute.manager [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Took 29.96 seconds to build instance. [ 843.839275] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.008548] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d3676-a15d-2057-10f7-af7cf91ca913, 'name': SearchDatastore_Task, 'duration_secs': 0.013563} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.008852] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.009099] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.009334] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.009486] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.009695] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.010014] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9907f97-1329-45f4-994b-679d70d2551a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.021176] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.021846] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.023761] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00751ae6-2d01-469f-86ab-3fb2e5d9a5e2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.029104] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 844.029104] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5213f9bf-3e55-122b-6746-e8f26b1db992" [ 844.029104] env[61006]: _type = "Task" [ 844.029104] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.044364] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5213f9bf-3e55-122b-6746-e8f26b1db992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.097907] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.218395] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.247597] env[61006]: DEBUG nova.scheduler.client.report [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.258184] env[61006]: DEBUG nova.compute.manager [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-vif-plugged-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.258184] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Acquiring lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.258389] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.258552] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.258723] env[61006]: DEBUG nova.compute.manager [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] No waiting events found dispatching network-vif-plugged-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.258882] env[61006]: WARNING nova.compute.manager [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received unexpected event network-vif-plugged-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c for instance with vm_state building and task_state spawning. [ 844.259178] env[61006]: DEBUG nova.compute.manager [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-changed-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.259540] env[61006]: DEBUG nova.compute.manager [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Refreshing instance network info cache due to event network-changed-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.260282] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Acquiring lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.260282] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Acquired lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.260282] env[61006]: DEBUG nova.network.neutron [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Refreshing network info cache for port 6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.309168] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0fb3dda5-bad2-4bc2-b9c0-b1c7b9314426 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.075s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.378033] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.378033] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.378033] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.378033] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.378242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.379403] env[61006]: INFO nova.compute.manager [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Terminating instance [ 844.381259] env[61006]: DEBUG nova.compute.manager [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.381473] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.382321] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262a47bb-69b2-4cf0-bb7f-aee0fba04b3a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.390541] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.390778] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da347787-d02c-418b-9e93-4f95ca5ec816 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.396831] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 844.396831] env[61006]: value = "task-1337038" [ 844.396831] env[61006]: _type = "Task" [ 844.396831] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.406336] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.543731] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5213f9bf-3e55-122b-6746-e8f26b1db992, 'name': SearchDatastore_Task, 'duration_secs': 0.015555} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.544685] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e2335eb-42ef-4f3e-acd2-1cc418a0e325 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.550193] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 844.550193] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d02285-6b68-f254-2277-de457c099543" [ 844.550193] env[61006]: _type = "Task" [ 844.550193] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.558404] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d02285-6b68-f254-2277-de457c099543, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.597334] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337037, 'name': ReconfigVM_Task, 'duration_secs': 1.004465} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.597811] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/0111452e-1b4f-499c-932d-f31364d1a14c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.598700] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71573351-c469-414c-b529-b1d1b9ba2d6f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.604640] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 844.604640] env[61006]: value = "task-1337039" [ 844.604640] env[61006]: _type = "Task" [ 844.604640] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.612712] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337039, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.720402] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.720700] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updated the network info_cache for instance {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 844.720961] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721171] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721333] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721531] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721690] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721838] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.721967] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 844.722324] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.753966] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.757243] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.578s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.757582] env[61006]: DEBUG nova.objects.instance [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lazy-loading 'resources' on Instance uuid 2a4089c9-4229-40bd-8d0e-706bba94655f {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.777394] env[61006]: INFO nova.scheduler.client.report [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Deleted allocations for instance ea7d5d49-ac76-4f2e-9456-912cf466fcc2 [ 844.796010] env[61006]: DEBUG nova.network.neutron [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.814756] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 844.849452] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.876090] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.876090] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.876359] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.876678] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.877031] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.877203] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.877544] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.877830] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.878149] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.878445] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.878984] env[61006]: DEBUG nova.virt.hardware [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.880115] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9567fcc3-2f9b-45d7-bcee-f08c7f291812 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.885373] env[61006]: DEBUG nova.network.neutron [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.894604] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97907bf7-19fc-479a-8071-5d2beb00dd3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.924451] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337038, 'name': PowerOffVM_Task, 'duration_secs': 0.239525} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.924451] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.924451] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 844.924964] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2a844d9-d349-4804-91ea-646843f1c028 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.996761] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 844.996761] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 844.996761] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Deleting the datastore file [datastore2] 28b83ea7-5433-472d-9e47-f73a4f2fb389 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.996761] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85068a1b-9b5b-4c23-889c-1dd85c19c81a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.003123] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for the task: (returnval){ [ 845.003123] env[61006]: value = "task-1337041" [ 845.003123] env[61006]: _type = "Task" [ 845.003123] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.012734] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.061106] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d02285-6b68-f254-2277-de457c099543, 'name': SearchDatastore_Task, 'duration_secs': 0.020216} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.061106] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.061106] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/aedaa5d6-e0f2-492c-a14b-3254863e1f06.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.062513] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cff54f0e-b3d2-4f6b-ad1a-f905ea6fefb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.067647] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 845.067647] env[61006]: value = "task-1337042" [ 845.067647] env[61006]: _type = "Task" [ 845.067647] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.077159] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.114717] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337039, 'name': Rename_Task, 'duration_secs': 0.142709} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.115019] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.115276] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8de5698d-80a5-48a4-a568-1837d77bccc8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.122775] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 845.122775] env[61006]: value = "task-1337043" [ 845.122775] env[61006]: _type = "Task" [ 845.122775] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.131013] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.225685] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.288977] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9db3355b-082a-41b9-a4f7-237e2f7e46af tempest-ImagesNegativeTestJSON-80570517 tempest-ImagesNegativeTestJSON-80570517-project-member] Lock "ea7d5d49-ac76-4f2e-9456-912cf466fcc2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.507s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.342535] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.388127] env[61006]: DEBUG oslo_concurrency.lockutils [req-891756a5-1747-4c93-9c75-c8368e24c594 req-72d30304-8625-4141-b75c-cdeddb02c116 service nova] Releasing lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.391483] env[61006]: DEBUG nova.compute.manager [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Received event network-vif-plugged-b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.391483] env[61006]: DEBUG oslo_concurrency.lockutils [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] Acquiring lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.391790] env[61006]: DEBUG oslo_concurrency.lockutils [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.391855] env[61006]: DEBUG oslo_concurrency.lockutils [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.391987] env[61006]: DEBUG nova.compute.manager [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] No waiting events found dispatching network-vif-plugged-b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.392589] env[61006]: WARNING nova.compute.manager [req-ffc1ccc5-c2f7-41f8-94e1-e89342446539 req-7a37e47d-10aa-456a-83cc-b4e58fe76583 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Received unexpected event network-vif-plugged-b597d7e4-0db4-40c4-90a2-f2245d7cdc67 for instance with vm_state building and task_state spawning. [ 845.464143] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Successfully updated port: b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.517845] env[61006]: DEBUG oslo_vmware.api [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Task: {'id': task-1337041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133271} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.517845] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.517845] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.518079] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.518357] env[61006]: INFO nova.compute.manager [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Took 1.14 seconds to destroy the instance on the hypervisor. [ 845.518813] env[61006]: DEBUG oslo.service.loopingcall [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.519350] env[61006]: DEBUG nova.compute.manager [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.520615] env[61006]: DEBUG nova.network.neutron [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 845.578549] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337042, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.619998] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd61fa90-6684-4691-965c-a9ed327b79ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.631705] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be52e09c-326b-4e9b-a1e1-fcc9eea8867b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.639620] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337043, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.675379] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8190c4ca-15ac-44ed-bfb4-55c9929e1f73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.682700] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6eb03d1-85d3-46e6-8a0d-f52edeb13ec7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.699904] env[61006]: DEBUG nova.compute.provider_tree [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.930205] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Successfully updated port: cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.968308] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.968455] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.968606] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 846.078771] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516985} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.079054] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/aedaa5d6-e0f2-492c-a14b-3254863e1f06.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.079280] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.079549] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55fc52ab-7fb7-41e4-916a-18f4e1caa700 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.086149] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 846.086149] env[61006]: value = "task-1337044" [ 846.086149] env[61006]: _type = "Task" [ 846.086149] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.093433] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337044, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.133433] env[61006]: DEBUG oslo_vmware.api [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337043, 'name': PowerOnVM_Task, 'duration_secs': 0.750696} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.133767] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.134034] env[61006]: INFO nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Took 8.68 seconds to spawn the instance on the hypervisor. [ 846.134254] env[61006]: DEBUG nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.135096] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26f6b6a-b56e-493f-b5bb-d6191c630526 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.203709] env[61006]: DEBUG nova.scheduler.client.report [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.286334] env[61006]: DEBUG nova.compute.manager [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-vif-plugged-cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.286561] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Acquiring lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.286774] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.286949] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.287138] env[61006]: DEBUG nova.compute.manager [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] No waiting events found dispatching network-vif-plugged-cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.287309] env[61006]: WARNING nova.compute.manager [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received unexpected event network-vif-plugged-cb74fd8b-90d3-477e-818c-51147a548d8e for instance with vm_state building and task_state spawning. [ 846.287473] env[61006]: DEBUG nova.compute.manager [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-changed-cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.287723] env[61006]: DEBUG nova.compute.manager [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Refreshing instance network info cache due to event network-changed-cb74fd8b-90d3-477e-818c-51147a548d8e. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 846.287923] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Acquiring lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.288074] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Acquired lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.288228] env[61006]: DEBUG nova.network.neutron [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Refreshing network info cache for port cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.358682] env[61006]: DEBUG nova.network.neutron [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.434073] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.509813] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.599802] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337044, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063365} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.599802] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.602107] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b545884-60f4-450c-be18-8f5b910da5c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.625078] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/aedaa5d6-e0f2-492c-a14b-3254863e1f06.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.625381] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38f09653-38f0-4fdc-b013-c36eab3ad1f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.649750] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 846.649750] env[61006]: value = "task-1337045" [ 846.649750] env[61006]: _type = "Task" [ 846.649750] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.656753] env[61006]: INFO nova.compute.manager [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Took 29.77 seconds to build instance. [ 846.665433] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.677644] env[61006]: DEBUG nova.network.neutron [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updating instance_info_cache with network_info: [{"id": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "address": "fa:16:3e:3d:5c:fc", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb597d7e4-0d", "ovs_interfaceid": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.707929] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.710190] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.904s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.712294] env[61006]: INFO nova.compute.claims [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.730187] env[61006]: INFO nova.scheduler.client.report [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Deleted allocations for instance 2a4089c9-4229-40bd-8d0e-706bba94655f [ 846.836429] env[61006]: DEBUG nova.network.neutron [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.861680] env[61006]: INFO nova.compute.manager [-] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Took 1.34 seconds to deallocate network for instance. [ 846.931089] env[61006]: DEBUG nova.network.neutron [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.158725] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52a4f221-9475-4aba-bedb-d1b5daf8cdff tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.237s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.163284] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337045, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.183628] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.183968] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Instance network_info: |[{"id": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "address": "fa:16:3e:3d:5c:fc", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb597d7e4-0d", "ovs_interfaceid": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 847.184422] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:5c:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b597d7e4-0db4-40c4-90a2-f2245d7cdc67', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.192345] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating folder: Project (931103a837fa4b2eb237dd4715ee0713). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.192656] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b4c8b5d-31b2-4193-afd9-b742dd350eea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.203028] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created folder: Project (931103a837fa4b2eb237dd4715ee0713) in parent group-v285275. [ 847.203028] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating folder: Instances. Parent ref: group-v285337. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.203219] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bca3c98-8c0f-4287-98ae-174c5780be8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.211620] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created folder: Instances in parent group-v285337. [ 847.211911] env[61006]: DEBUG oslo.service.loopingcall [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.212481] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 847.212569] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c426b97d-3b0b-4c57-ab74-027cd46808bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.236042] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.236042] env[61006]: value = "task-1337048" [ 847.236042] env[61006]: _type = "Task" [ 847.236042] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.242565] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7e6bdbc-8e56-4259-8542-cbf6416e2270 tempest-ServerAddressesNegativeTestJSON-1605873517 tempest-ServerAddressesNegativeTestJSON-1605873517-project-member] Lock "2a4089c9-4229-40bd-8d0e-706bba94655f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.187s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.248043] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337048, 'name': CreateVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.369048] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.433923] env[61006]: DEBUG oslo_concurrency.lockutils [req-29bfc057-ca9b-4a86-8395-0a87cb4acdd8 req-0415e877-61d4-4c74-84e2-4090c1f69314 service nova] Releasing lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.434613] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.434768] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.461031] env[61006]: DEBUG nova.compute.manager [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Received event network-changed-b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.461031] env[61006]: DEBUG nova.compute.manager [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Refreshing instance network info cache due to event network-changed-b597d7e4-0db4-40c4-90a2-f2245d7cdc67. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.461031] env[61006]: DEBUG oslo_concurrency.lockutils [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] Acquiring lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.461031] env[61006]: DEBUG oslo_concurrency.lockutils [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] Acquired lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.461031] env[61006]: DEBUG nova.network.neutron [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Refreshing network info cache for port b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.662150] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337045, 'name': ReconfigVM_Task, 'duration_secs': 0.827543} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.662461] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Reconfigured VM instance instance-00000047 to attach disk [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/aedaa5d6-e0f2-492c-a14b-3254863e1f06.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.663127] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7c16703-2ade-4fba-9abe-78f5a339a443 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.669717] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 847.669717] env[61006]: value = "task-1337049" [ 847.669717] env[61006]: _type = "Task" [ 847.669717] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.679890] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337049, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.744442] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337048, 'name': CreateVM_Task, 'duration_secs': 0.311972} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.744704] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.745461] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.745799] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.746024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.746314] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89772992-a03f-4f21-bff4-04e767cb4d3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.750857] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 847.750857] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf9e4c-81ca-e023-7c4b-1b986097fb60" [ 847.750857] env[61006]: _type = "Task" [ 847.750857] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.760981] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf9e4c-81ca-e023-7c4b-1b986097fb60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.978939] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.990019] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202771a5-d086-49f6-834f-0b2fd1391374 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.999722] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf73314-4911-4af7-b256-b7e8d9efe8b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.032705] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ba2c95-f94d-432c-9798-14e8f4ae1014 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.041186] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbbf3da3-2df4-4979-85da-7c7dee2ebdff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.057747] env[61006]: DEBUG nova.compute.provider_tree [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.179586] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337049, 'name': Rename_Task, 'duration_secs': 0.398484} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.180169] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.180477] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a016cf7-11ca-46e4-b0f1-0f56aa04d9ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.188833] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 848.188833] env[61006]: value = "task-1337050" [ 848.188833] env[61006]: _type = "Task" [ 848.188833] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.199999] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.260792] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf9e4c-81ca-e023-7c4b-1b986097fb60, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.261113] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.261344] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.261643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.261787] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.261987] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.262270] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-680f6764-c929-42a1-af07-7f3c3b4b5276 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.270120] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.270231] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.270939] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d05c6b5-47d8-4f86-b64c-1883015f4013 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.275841] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 848.275841] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52776fd8-ba25-aceb-cfd0-57fe29db3491" [ 848.275841] env[61006]: _type = "Task" [ 848.275841] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.280282] env[61006]: DEBUG nova.network.neutron [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Updating instance_info_cache with network_info: [{"id": "6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c", "address": "fa:16:3e:08:d6:0b", "network": {"id": "275ed65d-646f-4fd4-8eb8-02e2a18dfc1f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-183002772", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.153", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4693570ceab84d1babfa6f8f24447fcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c6c1a89-7c", "ovs_interfaceid": "6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cb74fd8b-90d3-477e-818c-51147a548d8e", "address": "fa:16:3e:00:9b:fc", "network": {"id": "e2aea23c-cb96-41ce-8490-93463b92c92e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1007169882", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.79", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4693570ceab84d1babfa6f8f24447fcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74fd8b-90", "ovs_interfaceid": "cb74fd8b-90d3-477e-818c-51147a548d8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.284865] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52776fd8-ba25-aceb-cfd0-57fe29db3491, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.562816] env[61006]: DEBUG nova.scheduler.client.report [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.582374] env[61006]: DEBUG nova.network.neutron [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updated VIF entry in instance network info cache for port b597d7e4-0db4-40c4-90a2-f2245d7cdc67. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.582797] env[61006]: DEBUG nova.network.neutron [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updating instance_info_cache with network_info: [{"id": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "address": "fa:16:3e:3d:5c:fc", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb597d7e4-0d", "ovs_interfaceid": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.698698] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337050, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.783197] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Releasing lock "refresh_cache-1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.783646] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance network_info: |[{"id": "6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c", "address": "fa:16:3e:08:d6:0b", "network": {"id": "275ed65d-646f-4fd4-8eb8-02e2a18dfc1f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-183002772", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.153", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4693570ceab84d1babfa6f8f24447fcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2c68e7-b690-42e2-9491-c3f9357cc66a", "external-id": "nsx-vlan-transportzone-321", "segmentation_id": 321, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c6c1a89-7c", "ovs_interfaceid": "6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cb74fd8b-90d3-477e-818c-51147a548d8e", "address": "fa:16:3e:00:9b:fc", "network": {"id": "e2aea23c-cb96-41ce-8490-93463b92c92e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1007169882", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.79", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4693570ceab84d1babfa6f8f24447fcd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74fd8b-90", "ovs_interfaceid": "cb74fd8b-90d3-477e-818c-51147a548d8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.784055] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:d6:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2c68e7-b690-42e2-9491-c3f9357cc66a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:9b:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb74fd8b-90d3-477e-818c-51147a548d8e', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.794460] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Creating folder: Project (4693570ceab84d1babfa6f8f24447fcd). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.795389] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d536d0a-6537-4c9b-9474-356c1e98a17f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.800842] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52776fd8-ba25-aceb-cfd0-57fe29db3491, 'name': SearchDatastore_Task, 'duration_secs': 0.016325} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.802131] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86709f90-90ca-4060-aad6-af3e304af110 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.807948] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 848.807948] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521a67ce-92e6-c820-7e78-d3596cd37c45" [ 848.807948] env[61006]: _type = "Task" [ 848.807948] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.809383] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Created folder: Project (4693570ceab84d1babfa6f8f24447fcd) in parent group-v285275. [ 848.809520] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Creating folder: Instances. Parent ref: group-v285340. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.812739] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bcfa133-226e-42de-aa38-d6c75b53b7b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.819912] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521a67ce-92e6-c820-7e78-d3596cd37c45, 'name': SearchDatastore_Task, 'duration_secs': 0.009604} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.820235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.820591] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 42061ea3-d1d1-4633-bd24-65f7ee302c1f/42061ea3-d1d1-4633-bd24-65f7ee302c1f.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.821802] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57deb24f-48b4-4961-af46-48477952610b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.823774] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Created folder: Instances in parent group-v285340. [ 848.824015] env[61006]: DEBUG oslo.service.loopingcall [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.824285] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.824798] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c14a1906-70e1-4c68-969f-a30afdea8555 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.842871] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 848.842871] env[61006]: value = "task-1337053" [ 848.842871] env[61006]: _type = "Task" [ 848.842871] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.847108] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.847108] env[61006]: value = "task-1337054" [ 848.847108] env[61006]: _type = "Task" [ 848.847108] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.854764] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.857450] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337054, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.069255] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.069804] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.072853] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.735s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.074399] env[61006]: INFO nova.compute.claims [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.086265] env[61006]: DEBUG oslo_concurrency.lockutils [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] Releasing lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.086680] env[61006]: DEBUG nova.compute.manager [req-de763b5e-9bd8-4974-b4af-a02bd85166b0 req-88d0fd32-2f5a-42c4-be8a-d76489020c6c service nova] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Received event network-vif-deleted-b6ca963a-897a-4678-b1ed-75c5a1099ee2 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.200040] env[61006]: DEBUG oslo_vmware.api [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337050, 'name': PowerOnVM_Task, 'duration_secs': 0.622248} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.200040] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.200040] env[61006]: INFO nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Took 9.30 seconds to spawn the instance on the hypervisor. [ 849.200351] env[61006]: DEBUG nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.201297] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20341145-945f-4c9d-b47f-8a88b0ed03c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.362476] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337054, 'name': CreateVM_Task, 'duration_secs': 0.429236} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.362980] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337053, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.363465] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.364522] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.365196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.365701] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.366153] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26f2264c-9ba5-416c-9baf-a05e87ece6b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.374122] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 849.374122] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ef962-123b-b065-e5d8-a3172c7a7a70" [ 849.374122] env[61006]: _type = "Task" [ 849.374122] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.382095] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ef962-123b-b065-e5d8-a3172c7a7a70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.582959] env[61006]: DEBUG nova.compute.utils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.586530] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.586717] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 849.680205] env[61006]: DEBUG nova.policy [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9abb864a5ff4b5fab9e64541aa3e419', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39c1f808ac524f4e974bab44ef759b4b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.718501] env[61006]: INFO nova.compute.manager [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Took 24.55 seconds to build instance. [ 849.854479] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611559} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.854761] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 42061ea3-d1d1-4633-bd24-65f7ee302c1f/42061ea3-d1d1-4633-bd24-65f7ee302c1f.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.854970] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.855242] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e97c1ce-03c4-4d61-8af7-de3946e8c34e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.861526] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 849.861526] env[61006]: value = "task-1337055" [ 849.861526] env[61006]: _type = "Task" [ 849.861526] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.869319] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337055, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.883242] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ef962-123b-b065-e5d8-a3172c7a7a70, 'name': SearchDatastore_Task, 'duration_secs': 0.033846} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.883242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.883242] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.883369] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.883507] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.883691] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.883976] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34f71867-b530-486d-a362-c6f160e4cc83 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.893140] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.893333] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 849.894087] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-512ed2ea-88bb-4407-80c8-5f7ad07b286d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.901218] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 849.901218] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52616188-6e49-57fb-4d56-02a6e90e49a0" [ 849.901218] env[61006]: _type = "Task" [ 849.901218] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.908978] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52616188-6e49-57fb-4d56-02a6e90e49a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.078763] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Successfully created port: 953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.090621] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.224499] env[61006]: DEBUG oslo_concurrency.lockutils [None req-15d5fd95-2c0a-4fc4-99f6-5d2092f286d1 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.554s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.357892] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1735b4-2320-4bc0-9cb4-b2b2e87de689 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.370110] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337055, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068508} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.371920] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.372723] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b02528b-7ef5-4025-ad27-e52b03f843a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.375676] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645dbda6-d15b-4742-983e-eaf9f34b47a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.397841] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 42061ea3-d1d1-4633-bd24-65f7ee302c1f/42061ea3-d1d1-4633-bd24-65f7ee302c1f.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.420644] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3410c3d4-3cae-46e3-b92f-fa9770d8f48a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.438952] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776c2f60-3de7-4114-9f65-43eede5a5788 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.447558] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52616188-6e49-57fb-4d56-02a6e90e49a0, 'name': SearchDatastore_Task, 'duration_secs': 0.015188} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.450726] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 850.450726] env[61006]: value = "task-1337056" [ 850.450726] env[61006]: _type = "Task" [ 850.450726] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.450933] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9804a5f4-635c-4e18-9b1e-8c1722ade88b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.453987] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83edae0-def5-4513-963d-ca0da1c6186e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.463369] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 850.463369] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5232435e-6582-782e-a140-cf4c2040f7f6" [ 850.463369] env[61006]: _type = "Task" [ 850.463369] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.474187] env[61006]: DEBUG nova.compute.provider_tree [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.475538] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.485480] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5232435e-6582-782e-a140-cf4c2040f7f6, 'name': SearchDatastore_Task, 'duration_secs': 0.013513} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.486315] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.486572] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a/1a7b5a77-b43e-4c96-ac18-36634a0e7d5a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 850.486822] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-014e0a58-ae8e-43a0-9926-cc8956d916e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.492804] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 850.492804] env[61006]: value = "task-1337057" [ 850.492804] env[61006]: _type = "Task" [ 850.492804] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.501641] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.529912] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.530532] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.531120] env[61006]: DEBUG nova.objects.instance [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'flavor' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.800413] env[61006]: INFO nova.compute.manager [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Rescuing [ 850.800686] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.800841] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.801016] env[61006]: DEBUG nova.network.neutron [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.968488] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337056, 'name': ReconfigVM_Task, 'duration_secs': 0.284221} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.968844] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 42061ea3-d1d1-4633-bd24-65f7ee302c1f/42061ea3-d1d1-4633-bd24-65f7ee302c1f.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.969540] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c81494a9-da1c-466a-b7d7-8fcb24b50a89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.977744] env[61006]: DEBUG nova.scheduler.client.report [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.981288] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 850.981288] env[61006]: value = "task-1337058" [ 850.981288] env[61006]: _type = "Task" [ 850.981288] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.992219] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337058, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.002299] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337057, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.035800] env[61006]: DEBUG nova.objects.instance [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'pci_requests' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.104047] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.130254] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.130400] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.130577] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.130849] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.131071] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.131289] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.131636] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.131875] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.132103] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.132348] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.132609] env[61006]: DEBUG nova.virt.hardware [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.133611] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c933f02e-1371-4f6d-b55b-d7b7cfcbe2b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.141757] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb1ef81-4c6d-404e-b355-5e151aa729c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.483173] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.483691] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.486347] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.723s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.487891] env[61006]: INFO nova.compute.claims [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.512259] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337058, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.512808] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628531} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.513393] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a/1a7b5a77-b43e-4c96-ac18-36634a0e7d5a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.516120] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.516120] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f3d21779-0f91-45c6-b23c-31c261baddba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.520604] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 851.520604] env[61006]: value = "task-1337059" [ 851.520604] env[61006]: _type = "Task" [ 851.520604] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.529615] env[61006]: DEBUG nova.compute.manager [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Received event network-vif-plugged-953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.529841] env[61006]: DEBUG oslo_concurrency.lockutils [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] Acquiring lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.530097] env[61006]: DEBUG oslo_concurrency.lockutils [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.530361] env[61006]: DEBUG oslo_concurrency.lockutils [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.530586] env[61006]: DEBUG nova.compute.manager [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] No waiting events found dispatching network-vif-plugged-953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.530859] env[61006]: WARNING nova.compute.manager [req-fb4b27b4-fed4-4859-8446-3681c475f855 req-f6cd6364-c8c1-4087-9015-2ab6ad5f46f3 service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Received unexpected event network-vif-plugged-953380b2-e816-4672-b9cd-2706604e550b for instance with vm_state building and task_state spawning. [ 851.535358] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337059, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.539353] env[61006]: DEBUG nova.objects.base [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 851.539353] env[61006]: DEBUG nova.network.neutron [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.542844] env[61006]: DEBUG nova.network.neutron [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Updating instance_info_cache with network_info: [{"id": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "address": "fa:16:3e:3f:75:e4", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63ee7dc4-4d", "ovs_interfaceid": "63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.600980] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Successfully updated port: 953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.677368] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b739b803-1f6e-49ab-ac5c-78af8b536471 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.147s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.995894] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337058, 'name': Rename_Task, 'duration_secs': 0.826758} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.998886] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.000699] env[61006]: DEBUG nova.compute.utils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.002563] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2f0f7ef-0cf2-45f7-a07a-eab972f67843 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.004526] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.004699] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 852.013029] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 852.013029] env[61006]: value = "task-1337060" [ 852.013029] env[61006]: _type = "Task" [ 852.013029] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.020219] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.029060] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337059, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065938} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.029323] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.030220] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa905e15-6e4c-434f-aec2-3868d431e7a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.046553] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-aedaa5d6-e0f2-492c-a14b-3254863e1f06" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.057901] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a/1a7b5a77-b43e-4c96-ac18-36634a0e7d5a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.059060] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b7f2a65-876f-484f-81e0-fe37abd9f5fd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.074882] env[61006]: DEBUG nova.policy [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cca6e2806aa45208ae618f6a78ccc0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe9eabeec9a941e68a9eae559e24ff4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.081283] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 852.081283] env[61006]: value = "task-1337061" [ 852.081283] env[61006]: _type = "Task" [ 852.081283] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.092775] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337061, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.104223] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.104223] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquired lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.104544] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.392304] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Successfully created port: f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.505083] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.526523] env[61006]: DEBUG oslo_vmware.api [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337060, 'name': PowerOnVM_Task, 'duration_secs': 0.511722} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.526807] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.527028] env[61006]: INFO nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Took 7.68 seconds to spawn the instance on the hypervisor. [ 852.527219] env[61006]: DEBUG nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.528051] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8785c0eb-36f4-49c5-9f1b-79ec813111b4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.593933] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337061, 'name': ReconfigVM_Task, 'duration_secs': 0.373814} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.597022] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.597358] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a/1a7b5a77-b43e-4c96-ac18-36634a0e7d5a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.598323] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5825db87-13c6-4b29-881d-3ff64f0fac33 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.600996] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bfdb68e-509b-407f-8e34-7a9fe5039798 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.608341] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 852.608341] env[61006]: value = "task-1337062" [ 852.608341] env[61006]: _type = "Task" [ 852.608341] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.609658] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 852.609658] env[61006]: value = "task-1337063" [ 852.609658] env[61006]: _type = "Task" [ 852.609658] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.626096] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.629433] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337063, 'name': Rename_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.642914] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.759019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.759019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.759019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.759019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.759019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.762376] env[61006]: INFO nova.compute.manager [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Terminating instance [ 852.765071] env[61006]: DEBUG nova.compute.manager [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.765431] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.766400] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3811fbf-cac5-4e80-8077-3d042e25e4de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.775808] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.779511] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9727d545-682a-4a5e-a1f5-cd96b60145ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.790808] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 852.790808] env[61006]: value = "task-1337064" [ 852.790808] env[61006]: _type = "Task" [ 852.790808] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.805501] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1337064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.823828] env[61006]: DEBUG nova.network.neutron [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Updating instance_info_cache with network_info: [{"id": "953380b2-e816-4672-b9cd-2706604e550b", "address": "fa:16:3e:77:cd:f1", "network": {"id": "ee3b5e79-2fdd-4e28-b896-75c5599c7a52", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1933383346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39c1f808ac524f4e974bab44ef759b4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap953380b2-e8", "ovs_interfaceid": "953380b2-e816-4672-b9cd-2706604e550b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.856729] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121e596f-02fc-4ac7-9fa7-a0ef985187af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.864711] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5c1ec8-fd18-4759-abf7-a0396aac2a85 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.896438] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c2fdea-bf21-46b9-a345-a6d694036e6f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.904327] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a3aa48-dd74-48fa-b166-d5d3e414a2c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.919861] env[61006]: DEBUG nova.compute.provider_tree [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.055682] env[61006]: INFO nova.compute.manager [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Took 23.12 seconds to build instance. [ 853.123165] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337062, 'name': PowerOffVM_Task, 'duration_secs': 0.232562} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.123858] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.124692] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd2fac2-82a4-4a8a-a2b8-f6ff3a960a5f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.130096] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337063, 'name': Rename_Task, 'duration_secs': 0.165576} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.130740] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.131090] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-faa422aa-3773-4160-8dca-300276b34ddf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.146243] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4928d5-209a-48e1-afc7-db3f4c96d61f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.150755] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 853.150755] env[61006]: value = "task-1337065" [ 853.150755] env[61006]: _type = "Task" [ 853.150755] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.165228] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337065, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.184173] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.184454] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87423e1e-aeeb-4ca8-8cde-43455594a2bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.190423] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 853.190423] env[61006]: value = "task-1337066" [ 853.190423] env[61006]: _type = "Task" [ 853.190423] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.199158] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337066, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.301118] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1337064, 'name': PowerOffVM_Task, 'duration_secs': 0.260135} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.301511] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.301728] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.302059] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dc24b26-fb65-44f2-929b-41c670336d1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.326540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Releasing lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.326924] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Instance network_info: |[{"id": "953380b2-e816-4672-b9cd-2706604e550b", "address": "fa:16:3e:77:cd:f1", "network": {"id": "ee3b5e79-2fdd-4e28-b896-75c5599c7a52", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1933383346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39c1f808ac524f4e974bab44ef759b4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap953380b2-e8", "ovs_interfaceid": "953380b2-e816-4672-b9cd-2706604e550b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.327382] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:cd:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e51ebca-e0f8-4b77-b155-4ff928eef130', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '953380b2-e816-4672-b9cd-2706604e550b', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.335054] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Creating folder: Project (39c1f808ac524f4e974bab44ef759b4b). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.335355] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d37cf8c7-cfde-4c7c-ad66-3b4a7a976843 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.345199] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Created folder: Project (39c1f808ac524f4e974bab44ef759b4b) in parent group-v285275. [ 853.345445] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Creating folder: Instances. Parent ref: group-v285343. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.345731] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bf0529e-df62-427e-bf3f-e12b326b710d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.358670] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Created folder: Instances in parent group-v285343. [ 853.358670] env[61006]: DEBUG oslo.service.loopingcall [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.358670] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.358670] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8da3d8bc-f1fa-4c5d-84a1-b3ddd00e6802 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.374331] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.374542] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.374721] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Deleting the datastore file [datastore2] 6e99894e-81b6-4a07-9ec7-caa16272b3ba {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.375406] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-732121c0-26b9-4528-8a30-839ec9fe1a3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.379516] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.379516] env[61006]: value = "task-1337070" [ 853.379516] env[61006]: _type = "Task" [ 853.379516] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.383858] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for the task: (returnval){ [ 853.383858] env[61006]: value = "task-1337071" [ 853.383858] env[61006]: _type = "Task" [ 853.383858] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.389886] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337070, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.394598] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1337071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.423673] env[61006]: DEBUG nova.scheduler.client.report [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.519765] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 853.549762] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.550020] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.550184] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.550370] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.550520] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.550673] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.550880] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.551057] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.551230] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.551392] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.551661] env[61006]: DEBUG nova.virt.hardware [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.552609] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba1c4a2-fef1-4d16-9468-051bb94f50ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.558198] env[61006]: DEBUG nova.compute.manager [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Received event network-changed-953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.558506] env[61006]: DEBUG nova.compute.manager [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Refreshing instance network info cache due to event network-changed-953380b2-e816-4672-b9cd-2706604e550b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.558717] env[61006]: DEBUG oslo_concurrency.lockutils [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] Acquiring lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.558912] env[61006]: DEBUG oslo_concurrency.lockutils [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] Acquired lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.559142] env[61006]: DEBUG nova.network.neutron [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Refreshing network info cache for port 953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.560947] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3271797d-2e92-4f72-ade3-a3708f1bffea tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.723s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.566426] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a94a2d0-4ff7-4530-bee7-e060d7ebf12e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.657676] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.658042] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.658410] env[61006]: DEBUG nova.objects.instance [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'flavor' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.665414] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337065, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.701351] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 853.701607] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.701859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.702019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.702204] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.702453] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1613fb71-4136-420d-968c-8aab428e77db {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.712600] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.712794] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.713544] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9c82232-951e-4996-a699-9d733fa1c2a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.718851] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 853.718851] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ac0271-6c10-1aa6-afec-dfc03d75fa9f" [ 853.718851] env[61006]: _type = "Task" [ 853.718851] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.726876] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ac0271-6c10-1aa6-afec-dfc03d75fa9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.892128] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337070, 'name': CreateVM_Task, 'duration_secs': 0.303749} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.893049] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.894529] env[61006]: DEBUG oslo_vmware.service [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fa1cd3-7dde-4c6e-901a-699e529707fb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.902018] env[61006]: DEBUG oslo_vmware.api [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Task: {'id': task-1337071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20353} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.902018] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.902134] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.902312] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.902529] env[61006]: INFO nova.compute.manager [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Took 1.14 seconds to destroy the instance on the hypervisor. [ 853.902815] env[61006]: DEBUG oslo.service.loopingcall [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.905027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.905027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.905154] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.905297] env[61006]: DEBUG nova.compute.manager [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.905396] env[61006]: DEBUG nova.network.neutron [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 853.906913] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44f7f8d8-c82f-4be7-9301-b1d086d7a71f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.911699] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 853.911699] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b48eb4-4957-a1a6-1d7f-4fcc5337572f" [ 853.911699] env[61006]: _type = "Task" [ 853.911699] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.918759] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b48eb4-4957-a1a6-1d7f-4fcc5337572f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.929817] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.930370] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.933745] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.914s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.934027] env[61006]: DEBUG nova.objects.instance [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lazy-loading 'resources' on Instance uuid 88e2bdc9-ab73-4e23-94b5-a45046835144 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.050318] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Successfully updated port: f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.166608] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337065, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.229280] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ac0271-6c10-1aa6-afec-dfc03d75fa9f, 'name': SearchDatastore_Task, 'duration_secs': 0.017258} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.233024] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0b0837f-389c-445a-ba8d-177a8a839061 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.235233] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 854.235233] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522c2c49-4ade-7917-e5af-06e5ee4d7fb2" [ 854.235233] env[61006]: _type = "Task" [ 854.235233] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.243160] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522c2c49-4ade-7917-e5af-06e5ee4d7fb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.272390] env[61006]: DEBUG nova.objects.instance [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'pci_requests' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.311942] env[61006]: DEBUG nova.network.neutron [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Updated VIF entry in instance network info cache for port 953380b2-e816-4672-b9cd-2706604e550b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.312461] env[61006]: DEBUG nova.network.neutron [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Updating instance_info_cache with network_info: [{"id": "953380b2-e816-4672-b9cd-2706604e550b", "address": "fa:16:3e:77:cd:f1", "network": {"id": "ee3b5e79-2fdd-4e28-b896-75c5599c7a52", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1933383346-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39c1f808ac524f4e974bab44ef759b4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e51ebca-e0f8-4b77-b155-4ff928eef130", "external-id": "nsx-vlan-transportzone-859", "segmentation_id": 859, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap953380b2-e8", "ovs_interfaceid": "953380b2-e816-4672-b9cd-2706604e550b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.422492] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.422880] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.423149] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.424144] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.424418] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.424688] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4c2b042-ad03-4d2e-85f4-fa5e3a5f2054 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.434366] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.434577] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.435377] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65322822-dcae-4bfd-a326-5977ec2d41f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.439193] env[61006]: DEBUG nova.compute.utils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.444066] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.444066] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.448971] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b69e62fc-dc26-42d0-9fef-bb85b38d5768 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.455436] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 854.455436] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5242456b-c6ad-9bbd-01aa-75bfb7efac89" [ 854.455436] env[61006]: _type = "Task" [ 854.455436] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.465625] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5242456b-c6ad-9bbd-01aa-75bfb7efac89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.490058] env[61006]: DEBUG nova.compute.manager [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Received event network-vif-plugged-f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.490242] env[61006]: DEBUG oslo_concurrency.lockutils [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.490815] env[61006]: DEBUG oslo_concurrency.lockutils [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.490815] env[61006]: DEBUG oslo_concurrency.lockutils [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.490815] env[61006]: DEBUG nova.compute.manager [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] No waiting events found dispatching network-vif-plugged-f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.491225] env[61006]: WARNING nova.compute.manager [req-a3277ecc-f009-47f4-afde-dd605e707e6f req-3f2cd310-8e80-45ad-88ea-78e2308d0b13 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Received unexpected event network-vif-plugged-f537f6f0-2a6c-4479-84ea-686716df3690 for instance with vm_state building and task_state spawning. [ 854.492828] env[61006]: DEBUG nova.policy [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8b0db2570144795b04809d82e643764', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ad8b37b118c4c8a8fde488ffdc44621', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.554632] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.554777] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.554925] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.667769] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337065, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.751554] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522c2c49-4ade-7917-e5af-06e5ee4d7fb2, 'name': SearchDatastore_Task, 'duration_secs': 0.020237} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.751554] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.751554] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. {{(pid=61006) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 854.751554] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff341a59-564c-4b1c-82c9-aaa774ad0eff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.758548] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 854.758548] env[61006]: value = "task-1337072" [ 854.758548] env[61006]: _type = "Task" [ 854.758548] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.769926] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337072, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.776834] env[61006]: DEBUG nova.objects.base [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 854.777096] env[61006]: DEBUG nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.790215] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e1765f-2ac2-44eb-858a-6bfeba5cdac3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.801875] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce03777-e4c6-441c-a896-85b245fb4269 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.836876] env[61006]: DEBUG oslo_concurrency.lockutils [req-12f434d9-7b36-4b0e-aadb-87a1fc3302c3 req-9d51bac9-2c5a-4ff3-9549-52a72f75d81b service nova] Releasing lock "refresh_cache-9c9fa347-bcfe-4009-af72-5f427e9d234a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.838515] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Successfully created port: 0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.840930] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1903ca8-3d1c-4baa-bc00-3e0460071724 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.846449] env[61006]: DEBUG nova.policy [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.851874] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cad394c-7cc0-424d-875b-b089eb9786f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.866209] env[61006]: DEBUG nova.compute.provider_tree [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.932595] env[61006]: DEBUG oslo_concurrency.lockutils [None req-67002090-297f-4d10-ae3a-c451aa12ffc6 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "509160cc-6df7-4f80-8d01-ebac3f4c6e2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.933263] env[61006]: DEBUG oslo_concurrency.lockutils [None req-67002090-297f-4d10-ae3a-c451aa12ffc6 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "509160cc-6df7-4f80-8d01-ebac3f4c6e2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.944880] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.968675] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 854.968948] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Creating directory with path [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.969244] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91f38f01-19df-4444-af03-1bbce646c6fa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.997057] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Created directory with path [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.997304] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Fetch image to [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 854.997492] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Downloading image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk on the data store datastore1 {{(pid=61006) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 854.998636] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a203d1a7-b234-4b44-9194-1f08519b6c35 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.009018] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc16326-4c08-4897-ac10-a455e553c0f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.021596] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bc2b5a-b1ba-41b2-9d77-6a4b4df5a360 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.054023] env[61006]: DEBUG nova.network.neutron [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.057902] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c66655-cf84-4791-a862-fe9c75bad554 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.068912] env[61006]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-84b73966-1618-44e8-abc9-edb6a6349fd3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.090909] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Downloading image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to the data store datastore1 {{(pid=61006) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 855.110378] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.167123] env[61006]: DEBUG oslo_vmware.rw_handles [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 855.169411] env[61006]: DEBUG oslo_vmware.api [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337065, 'name': PowerOnVM_Task, 'duration_secs': 1.539553} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.170570] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.170807] env[61006]: INFO nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Took 12.86 seconds to spawn the instance on the hypervisor. [ 855.171021] env[61006]: DEBUG nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.235018] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4a456d-8a27-47a3-a8bf-40060d8dfe10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.272328] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337072, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.322806] env[61006]: DEBUG nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Successfully created port: 0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.370173] env[61006]: DEBUG nova.scheduler.client.report [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.385241] env[61006]: DEBUG nova.network.neutron [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.435945] env[61006]: DEBUG nova.compute.manager [None req-67002090-297f-4d10-ae3a-c451aa12ffc6 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 509160cc-6df7-4f80-8d01-ebac3f4c6e2d] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.561194] env[61006]: INFO nova.compute.manager [-] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Took 1.66 seconds to deallocate network for instance. [ 855.772775] env[61006]: INFO nova.compute.manager [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Took 28.31 seconds to build instance. [ 855.780139] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337072, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552038} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.780438] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. [ 855.781313] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8c7f06-167a-49f8-98cf-72f544100e19 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.807520] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.808222] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76794f8c-95c9-4516-9787-7a0911fbb00d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.827286] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 855.827286] env[61006]: value = "task-1337073" [ 855.827286] env[61006]: _type = "Task" [ 855.827286] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.835860] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337073, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.875476] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.877864] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.483s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.878273] env[61006]: DEBUG nova.objects.instance [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lazy-loading 'resources' on Instance uuid 598c0d72-d679-49a8-b17c-f5f341c205e8 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.887500] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.887907] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Instance network_info: |[{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.888822] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:f0:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb68953b-dee5-4d9d-b47b-277336ba76dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f537f6f0-2a6c-4479-84ea-686716df3690', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.897196] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Creating folder: Project (fe9eabeec9a941e68a9eae559e24ff4c). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.897884] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb8e75c8-dcea-4279-ad6a-e40b96f590db {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.901009] env[61006]: INFO nova.scheduler.client.report [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Deleted allocations for instance 88e2bdc9-ab73-4e23-94b5-a45046835144 [ 855.915360] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Created folder: Project (fe9eabeec9a941e68a9eae559e24ff4c) in parent group-v285275. [ 855.915567] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Creating folder: Instances. Parent ref: group-v285346. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.916608] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-373f566f-c588-44c1-94df-f7a51440c6e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.926999] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Created folder: Instances in parent group-v285346. [ 855.927321] env[61006]: DEBUG oslo.service.loopingcall [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.927532] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.927754] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aeb335a1-0632-4699-9875-f1e0c260f225 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.943480] env[61006]: DEBUG nova.compute.manager [None req-67002090-297f-4d10-ae3a-c451aa12ffc6 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 509160cc-6df7-4f80-8d01-ebac3f4c6e2d] Instance disappeared before build. {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 855.949592] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.949592] env[61006]: value = "task-1337076" [ 855.949592] env[61006]: _type = "Task" [ 855.949592] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.956731] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.965402] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337076, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.987340] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.987863] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.987863] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.988068] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.988401] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.988566] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.988793] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.989079] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.989260] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.989440] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.989645] env[61006]: DEBUG nova.virt.hardware [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.990608] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14690ab8-87c2-43c2-b306-cc8274feb96d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.001948] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fbbbfe-7a9f-4785-8f51-8e5029d741cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.074060] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.092775] env[61006]: DEBUG oslo_vmware.rw_handles [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 856.093102] env[61006]: DEBUG oslo_vmware.rw_handles [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 856.202128] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Downloaded image file data 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk on the data store datastore1 {{(pid=61006) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 856.204610] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 856.204884] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Copying Virtual Disk [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk to [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.205479] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74c5306e-7f9f-495f-b9a6-588797c9611d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.213280] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 856.213280] env[61006]: value = "task-1337077" [ 856.213280] env[61006]: _type = "Task" [ 856.213280] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.223229] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.275733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a7a0b956-1bce-45b5-846f-7bd45b9a37d9 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.944s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.316946] env[61006]: DEBUG nova.compute.manager [req-23d4e887-d69c-4e94-b640-956ef91958ae req-5512cd96-4e56-4394-869f-f45dc7056265 service nova] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Received event network-vif-deleted-9e1bc75a-8bdf-4cfb-ba15-044d90816855 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.339458] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337073, 'name': ReconfigVM_Task, 'duration_secs': 0.307152} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.339888] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Reconfigured VM instance instance-00000047 to attach disk [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.340827] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310c6c9c-4cb6-41ed-b98e-127fad31abf6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.369449] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7670f97-cbc5-4c90-8261-f4913599fb4d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.389925] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 856.389925] env[61006]: value = "task-1337078" [ 856.389925] env[61006]: _type = "Task" [ 856.389925] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.399816] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.412512] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c5638b83-4841-4ba8-9bde-4719155f3f27 tempest-SecurityGroupsTestJSON-1960792207 tempest-SecurityGroupsTestJSON-1960792207-project-member] Lock "88e2bdc9-ab73-4e23-94b5-a45046835144" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.556s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.459916] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337076, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.461035] env[61006]: DEBUG oslo_concurrency.lockutils [None req-67002090-297f-4d10-ae3a-c451aa12ffc6 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "509160cc-6df7-4f80-8d01-ebac3f4c6e2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.528s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.470882] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Successfully updated port: 0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.677321] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8ebb45-4a25-4cc3-bc27-a13ea5733fb8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.685207] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f6dcf5-0bab-4d8b-8e87-482ce76509f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.719280] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1066f4-a4c9-498a-b2fd-aaab979c7efb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.729369] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337077, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.732755] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745e4d98-b45c-4f5e-a96a-7ad8d9b8f9a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.747411] env[61006]: DEBUG nova.compute.provider_tree [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.900111] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337078, 'name': ReconfigVM_Task, 'duration_secs': 0.207773} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.900411] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.900674] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5558593-4c62-402c-82df-d8a9865cfc41 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.906988] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 856.906988] env[61006]: value = "task-1337079" [ 856.906988] env[61006]: _type = "Task" [ 856.906988] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.914986] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.959599] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337076, 'name': CreateVM_Task, 'duration_secs': 0.513375} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.959776] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.960511] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.960675] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.960995] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.961266] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9649e495-508a-4f99-b9fe-9ff3159f67dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.965806] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 856.965806] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521ec646-1c4f-6e1e-6b87-ab3accb65344" [ 856.965806] env[61006]: _type = "Task" [ 856.965806] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.973654] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.973834] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.973964] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.975032] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521ec646-1c4f-6e1e-6b87-ab3accb65344, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.092386] env[61006]: DEBUG nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Successfully updated port: 0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.107939] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Received event network-changed-f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.108268] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Refreshing instance network info cache due to event network-changed-f537f6f0-2a6c-4479-84ea-686716df3690. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.108509] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Acquiring lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.108676] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Acquired lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.108812] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Refreshing network info cache for port f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.225298] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337077, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.769146} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.225596] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Copied Virtual Disk [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk to [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.225738] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleting the datastore file [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/tmp-sparse.vmdk {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.225997] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cb5625e-c57d-4f64-a7f8-2656368f498a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.232291] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 857.232291] env[61006]: value = "task-1337080" [ 857.232291] env[61006]: _type = "Task" [ 857.232291] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.240230] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.251139] env[61006]: DEBUG nova.scheduler.client.report [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.418034] env[61006]: DEBUG oslo_vmware.api [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337079, 'name': PowerOnVM_Task, 'duration_secs': 0.437359} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.418034] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.421225] env[61006]: DEBUG nova.compute.manager [None req-bc5ea404-46b2-4f4d-8644-21ed8930645a tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.422436] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee04148-0d16-42ff-9d17-f9f988d24f67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.480046] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.480952] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.480952] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.509324] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.594761] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.594963] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.595180] env[61006]: DEBUG nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.715100] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d1e09ace-177d-430a-9e65-d01368815272" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.716999] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.725726] env[61006]: DEBUG nova.network.neutron [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating instance_info_cache with network_info: [{"id": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "address": "fa:16:3e:c0:e7:6d", "network": {"id": "3d20b60a-ee54-49f6-80e6-835258e52eac", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-743242808-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ad8b37b118c4c8a8fde488ffdc44621", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e26f4ec-25", "ovs_interfaceid": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.743761] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.046086} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.746715] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.746937] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Moving file from [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 to [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0. {{(pid=61006) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 857.747453] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4d18813b-76fc-49fd-b565-6615dfcbe44d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.756686] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.758991] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 857.758991] env[61006]: value = "task-1337081" [ 857.758991] env[61006]: _type = "Task" [ 857.758991] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.759189] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.250s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.759383] env[61006]: DEBUG nova.objects.instance [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lazy-loading 'resources' on Instance uuid 29c41817-2189-4622-8f35-86f61eb34bed {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.777161] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337081, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.782086] env[61006]: INFO nova.scheduler.client.report [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted allocations for instance 598c0d72-d679-49a8-b17c-f5f341c205e8 [ 857.856652] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updated VIF entry in instance network info cache for port f537f6f0-2a6c-4479-84ea-686716df3690. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.857057] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.130542] env[61006]: WARNING nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] 98dcba29-9f32-40eb-b526-46e091c2a330 already exists in list: networks containing: ['98dcba29-9f32-40eb-b526-46e091c2a330']. ignoring it [ 858.221421] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.228733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Releasing lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.229041] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Instance network_info: |[{"id": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "address": "fa:16:3e:c0:e7:6d", "network": {"id": "3d20b60a-ee54-49f6-80e6-835258e52eac", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-743242808-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ad8b37b118c4c8a8fde488ffdc44621", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e26f4ec-25", "ovs_interfaceid": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.229424] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:e7:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4aa1eda7-48b9-4fa2-af0b-94c718313af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e26f4ec-25ca-4a63-8c6a-b6767eda7456', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.237148] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Creating folder: Project (6ad8b37b118c4c8a8fde488ffdc44621). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.237488] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2576754-a469-47ab-a192-1919edd1cbac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.251289] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Created folder: Project (6ad8b37b118c4c8a8fde488ffdc44621) in parent group-v285275. [ 858.251637] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Creating folder: Instances. Parent ref: group-v285349. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.251909] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cdd261e8-4650-42a2-b70c-bd7dc266560b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.264246] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Created folder: Instances in parent group-v285349. [ 858.264554] env[61006]: DEBUG oslo.service.loopingcall [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.265204] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.268567] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf473801-f56f-465f-92cf-49e9822ef0cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.289111] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.289365] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.289575] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.289747] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.289913] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.294439] env[61006]: INFO nova.compute.manager [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Terminating instance [ 858.298079] env[61006]: DEBUG nova.compute.manager [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 858.298079] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.298715] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7333b28-c0e8-4bd9-9b08-f63308827256 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.307181] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.307181] env[61006]: value = "task-1337084" [ 858.307181] env[61006]: _type = "Task" [ 858.307181] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.307462] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337081, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.046669} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.308407] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] File moved {{(pid=61006) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 858.308619] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Cleaning up location [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 858.308785] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleting the datastore file [datastore1] vmware_temp/91fa8ef2-0bed-4358-a608-c06e098e2d3c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.312216] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3392fa7-95bf-4d3f-a78f-0c39699349d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.318953] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.322336] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba9a6b69-64cb-41cd-b710-8ab564a1733a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.327630] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337084, 'name': CreateVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.329513] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 858.329513] env[61006]: value = "task-1337085" [ 858.329513] env[61006]: _type = "Task" [ 858.329513] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.335981] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 858.335981] env[61006]: value = "task-1337086" [ 858.335981] env[61006]: _type = "Task" [ 858.335981] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.345818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5c6fc042-4305-462a-836f-cd39b0cc7461 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "598c0d72-d679-49a8-b17c-f5f341c205e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.099s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.350583] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.360422] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Releasing lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.361982] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Received event network-vif-plugged-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.361982] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.361982] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.361982] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.361982] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] No waiting events found dispatching network-vif-plugged-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.361982] env[61006]: WARNING nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Received unexpected event network-vif-plugged-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 for instance with vm_state building and task_state spawning. [ 858.362186] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Received event network-changed-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.363785] env[61006]: DEBUG nova.compute.manager [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Refreshing instance network info cache due to event network-changed-0e26f4ec-25ca-4a63-8c6a-b6767eda7456. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.363785] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Acquiring lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.363785] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Acquired lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.363785] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Refreshing network info cache for port 0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.364030] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.464246] env[61006]: DEBUG nova.network.neutron [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "address": "fa:16:3e:2b:cd:e9", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0df9528c-28", "ovs_interfaceid": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.581433] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426cb885-57e2-497c-96bb-801de4a7af26 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.589997] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a14523-b728-42fd-8c81-219387da520e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.620717] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb80818b-5908-48b2-a5fe-aad5da41b584 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.628683] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2adc3c3-96f7-4cd3-bfb3-01dbd3e8d96d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.644145] env[61006]: DEBUG nova.compute.provider_tree [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.674363] env[61006]: DEBUG nova.compute.manager [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-plugged-0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.674585] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.674790] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.674958] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.675172] env[61006]: DEBUG nova.compute.manager [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] No waiting events found dispatching network-vif-plugged-0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.675345] env[61006]: WARNING nova.compute.manager [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received unexpected event network-vif-plugged-0df9528c-287c-45f7-8d8d-b43dc00c0b24 for instance with vm_state active and task_state None. [ 858.675504] env[61006]: DEBUG nova.compute.manager [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-changed-0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.675657] env[61006]: DEBUG nova.compute.manager [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing instance network info cache due to event network-changed-0df9528c-287c-45f7-8d8d-b43dc00c0b24. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.675816] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.742483] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.819101] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337084, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.844798] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.053643} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.845079] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.846198] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bba07d75-27c3-4aca-b14f-dd058d193127 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.851856] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337086, 'name': PowerOffVM_Task, 'duration_secs': 0.26058} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.852570] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.853022] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.853097] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2680cfb0-ad43-4fbb-8132-40ed517a3d4c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.855951] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 858.855951] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efe2e9-8f32-d8c1-2e87-5826058943b7" [ 858.855951] env[61006]: _type = "Task" [ 858.855951] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.865014] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efe2e9-8f32-d8c1-2e87-5826058943b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.969954] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.970811] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.970811] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.970940] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.971059] env[61006]: DEBUG nova.network.neutron [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing network info cache for port 0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.973360] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232f40c1-7bd1-45e1-a390-10c974f10463 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.977846] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.978121] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.978326] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Deleting the datastore file [datastore2] 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.979014] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-664230bd-e28b-4080-9687-dba21a8e0158 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.993370] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.993612] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.993774] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.993986] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.994156] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.994307] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.994511] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.994671] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.994837] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.995068] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.995190] env[61006]: DEBUG nova.virt.hardware [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.001862] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfiguring VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 859.005417] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dcbc527-94d6-4dc9-8027-da615ff56c97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.017338] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for the task: (returnval){ [ 859.017338] env[61006]: value = "task-1337088" [ 859.017338] env[61006]: _type = "Task" [ 859.017338] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.024634] env[61006]: DEBUG oslo_vmware.api [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 859.024634] env[61006]: value = "task-1337089" [ 859.024634] env[61006]: _type = "Task" [ 859.024634] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.028062] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.036853] env[61006]: DEBUG oslo_vmware.api [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337089, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.114967] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updated VIF entry in instance network info cache for port 0e26f4ec-25ca-4a63-8c6a-b6767eda7456. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.115336] env[61006]: DEBUG nova.network.neutron [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating instance_info_cache with network_info: [{"id": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "address": "fa:16:3e:c0:e7:6d", "network": {"id": "3d20b60a-ee54-49f6-80e6-835258e52eac", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-743242808-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ad8b37b118c4c8a8fde488ffdc44621", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e26f4ec-25", "ovs_interfaceid": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.147007] env[61006]: DEBUG nova.scheduler.client.report [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.319995] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337084, 'name': CreateVM_Task, 'duration_secs': 0.570117} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.320343] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.321076] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.321262] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.321701] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.321927] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb9f37e2-4e4c-4f45-968e-a93ec18c2a0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.327425] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 859.327425] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5282b1b0-f9e1-999b-4bea-5131f6bb2ca7" [ 859.327425] env[61006]: _type = "Task" [ 859.327425] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.339115] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5282b1b0-f9e1-999b-4bea-5131f6bb2ca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.366621] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efe2e9-8f32-d8c1-2e87-5826058943b7, 'name': SearchDatastore_Task, 'duration_secs': 0.010494} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.366881] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.367159] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 9c9fa347-bcfe-4009-af72-5f427e9d234a/9c9fa347-bcfe-4009-af72-5f427e9d234a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.367472] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.367669] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.367887] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-619896fc-c30f-415c-b226-3c644e05775c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.370353] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5de4ea7b-7eec-409b-b793-85769ffdc64b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.377612] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 859.377612] env[61006]: value = "task-1337090" [ 859.377612] env[61006]: _type = "Task" [ 859.377612] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.381858] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.382040] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.383084] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c69f9d69-656b-4791-a04e-1299abf1fdb6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.388105] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.391795] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 859.391795] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52179ce3-4c33-2938-5d8f-f0face87f3d8" [ 859.391795] env[61006]: _type = "Task" [ 859.391795] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.400448] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52179ce3-4c33-2938-5d8f-f0face87f3d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.529135] env[61006]: DEBUG oslo_vmware.api [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Task: {'id': task-1337088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326028} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.532974] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.533257] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.533491] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.533689] env[61006]: INFO nova.compute.manager [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Took 1.24 seconds to destroy the instance on the hypervisor. [ 859.534170] env[61006]: DEBUG oslo.service.loopingcall [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.534464] env[61006]: DEBUG nova.compute.manager [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 859.534580] env[61006]: DEBUG nova.network.neutron [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 859.543367] env[61006]: DEBUG oslo_vmware.api [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337089, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.617862] env[61006]: DEBUG oslo_concurrency.lockutils [req-ba6d37c4-6662-4ed2-9f5f-f819c9cf234e req-5ae0fec2-3fb7-42f1-9cdd-b5bb0d38eee7 service nova] Releasing lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.652742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.656713] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.898s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.658625] env[61006]: INFO nova.compute.claims [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.694109] env[61006]: INFO nova.scheduler.client.report [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted allocations for instance 29c41817-2189-4622-8f35-86f61eb34bed [ 859.719709] env[61006]: DEBUG nova.network.neutron [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updated VIF entry in instance network info cache for port 0df9528c-287c-45f7-8d8d-b43dc00c0b24. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.720189] env[61006]: DEBUG nova.network.neutron [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "address": "fa:16:3e:2b:cd:e9", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0df9528c-28", "ovs_interfaceid": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.845283] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5282b1b0-f9e1-999b-4bea-5131f6bb2ca7, 'name': SearchDatastore_Task, 'duration_secs': 0.032289} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.845728] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.846063] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.846352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.889885] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337090, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.906016] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52179ce3-4c33-2938-5d8f-f0face87f3d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009002} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.907280] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96f0e609-339f-4125-9b08-587b9353e286 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.914144] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 859.914144] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52945f11-6dfc-14b2-d9ac-1781c92413d3" [ 859.914144] env[61006]: _type = "Task" [ 859.914144] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.926130] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52945f11-6dfc-14b2-d9ac-1781c92413d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.038706] env[61006]: DEBUG oslo_vmware.api [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337089, 'name': ReconfigVM_Task, 'duration_secs': 0.652853} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.039258] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.039474] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfigured VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 860.205700] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a6a9497-b53d-4d35-bd8a-a2d3a3f9117f tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "29c41817-2189-4622-8f35-86f61eb34bed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.850s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.225355] env[61006]: DEBUG oslo_concurrency.lockutils [req-e68be382-4193-4ed0-929b-b17359ed9dfb req-64689c47-b39e-4599-af91-8239c1a01c68 service nova] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.389522] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.389810] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 9c9fa347-bcfe-4009-af72-5f427e9d234a/9c9fa347-bcfe-4009-af72-5f427e9d234a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.390868] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.390868] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22472588-25bf-46bd-a03a-d401b6bf834f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.398481] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 860.398481] env[61006]: value = "task-1337091" [ 860.398481] env[61006]: _type = "Task" [ 860.398481] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.407800] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.427175] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52945f11-6dfc-14b2-d9ac-1781c92413d3, 'name': SearchDatastore_Task, 'duration_secs': 0.041425} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.427618] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.428138] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.428476] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.428720] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.428975] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89ead605-321d-496f-9f23-c0398ecb8037 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.431278] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b47e7ca-8824-49d0-ae38-054939c67223 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.440558] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 860.440558] env[61006]: value = "task-1337092" [ 860.440558] env[61006]: _type = "Task" [ 860.440558] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.445820] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.446017] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.447128] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17aad64a-bf8e-447c-a39d-a5703b9df8f9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.453224] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337092, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.456253] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 860.456253] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c475ca-79e8-328a-ced4-acf6de420a27" [ 860.456253] env[61006]: _type = "Task" [ 860.456253] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.466184] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c475ca-79e8-328a-ced4-acf6de420a27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.544373] env[61006]: DEBUG oslo_concurrency.lockutils [None req-620e163a-6e04-4bd6-a362-614eb455c544 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.886s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.577068] env[61006]: DEBUG nova.network.neutron [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.681686] env[61006]: INFO nova.compute.manager [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Rescuing [ 860.682112] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.682338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.682578] env[61006]: DEBUG nova.network.neutron [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.912023] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078797} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.912308] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.913428] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad323f68-321c-4358-b4b5-414e2fb178e4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.917512] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14207e42-2120-41f9-b615-b29f34dad601 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.934759] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3758a9b-e49b-49b2-b47d-29db5f9b9c93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.946522] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 9c9fa347-bcfe-4009-af72-5f427e9d234a/9c9fa347-bcfe-4009-af72-5f427e9d234a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.946917] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bbf536a-01ee-4b86-bfc0-251aec15967d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.999822] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8290b818-9f4d-4282-a761-6086e4ebad8d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.002981] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 861.002981] env[61006]: value = "task-1337093" [ 861.002981] env[61006]: _type = "Task" [ 861.002981] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.009901] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337092, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.010198] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c475ca-79e8-328a-ced4-acf6de420a27, 'name': SearchDatastore_Task, 'duration_secs': 0.037964} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.017146] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-648b3cf4-5d78-483d-88ed-8df35cf058bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.021585] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ccdf84-00ee-46f7-9919-d0a8059d1187 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.029971] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337093, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.035284] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 861.035284] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5233b970-9fed-4163-6e93-b66b313cb404" [ 861.035284] env[61006]: _type = "Task" [ 861.035284] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.045350] env[61006]: DEBUG nova.compute.provider_tree [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.059968] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5233b970-9fed-4163-6e93-b66b313cb404, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.079573] env[61006]: INFO nova.compute.manager [-] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Took 1.54 seconds to deallocate network for instance. [ 861.173279] env[61006]: DEBUG nova.compute.manager [req-58913b43-9119-43a9-8380-5ad243e712b6 req-09d81dac-b231-4b28-b88e-8308d8342a2b service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-vif-deleted-cb74fd8b-90d3-477e-818c-51147a548d8e {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.173279] env[61006]: DEBUG nova.compute.manager [req-58913b43-9119-43a9-8380-5ad243e712b6 req-09d81dac-b231-4b28-b88e-8308d8342a2b service nova] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Received event network-vif-deleted-6c6c1a89-7cfa-44d2-bb26-d3b4c8846a2c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.460530] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337092, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.929813} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.460530] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.460530] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.460530] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92ae82aa-947e-4fa4-ad81-30faca34e5c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.470347] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 861.470347] env[61006]: value = "task-1337094" [ 861.470347] env[61006]: _type = "Task" [ 861.470347] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.480013] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337094, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.485124] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "84befce1-ccd4-4937-9656-e5cbb7882f47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.485367] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.524686] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337093, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.525762] env[61006]: DEBUG nova.network.neutron [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [{"id": "ba96f533-658c-49f4-a9a4-48e129d5c197", "address": "fa:16:3e:bb:35:70", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba96f533-65", "ovs_interfaceid": "ba96f533-658c-49f4-a9a4-48e129d5c197", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.527297] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.527530] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.550808] env[61006]: DEBUG nova.scheduler.client.report [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.561227] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5233b970-9fed-4163-6e93-b66b313cb404, 'name': SearchDatastore_Task, 'duration_secs': 0.045822} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.562057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.562256] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] d174a8ec-867e-4fea-b878-2a9af1476949/d174a8ec-867e-4fea-b878-2a9af1476949.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.562466] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2161782b-2e92-4965-b539-586d15b359f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.569815] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 861.569815] env[61006]: value = "task-1337095" [ 861.569815] env[61006]: _type = "Task" [ 861.569815] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.579774] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.585901] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.980293] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337094, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074456} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.980567] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.981379] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa4f244-9acd-4dd4-85a3-3d34d0afb9a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.996047] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.007558] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.008134] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3be6db92-9420-441a-80f2-d1392b5853bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.032656] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.035447] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.037886] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337093, 'name': ReconfigVM_Task, 'duration_secs': 0.605734} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.039875] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 9c9fa347-bcfe-4009-af72-5f427e9d234a/9c9fa347-bcfe-4009-af72-5f427e9d234a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.040590] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 862.040590] env[61006]: value = "task-1337096" [ 862.040590] env[61006]: _type = "Task" [ 862.040590] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.040791] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61961f19-75d7-4540-8fed-33293d3251df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.052940] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.054371] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 862.054371] env[61006]: value = "task-1337097" [ 862.054371] env[61006]: _type = "Task" [ 862.054371] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.064293] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.064752] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.073309] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.019s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.074843] env[61006]: INFO nova.compute.claims [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.078389] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337097, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.089016] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.445505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-385eff2a-3efb-49b9-8063-c01a39357798" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.445950] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-385eff2a-3efb-49b9-8063-c01a39357798" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.446422] env[61006]: DEBUG nova.objects.instance [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'flavor' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.534197] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.558238] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.559886] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.570952] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337097, 'name': Rename_Task, 'duration_secs': 0.185658} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.571991] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.571991] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ffa28cc1-2574-48ce-b2d7-1190eb601c25 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.577070] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.577070] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7533e2fe-79b2-41af-8a35-e1dbd3e13c6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.586141] env[61006]: DEBUG nova.compute.utils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.588199] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 862.588199] env[61006]: value = "task-1337098" [ 862.588199] env[61006]: _type = "Task" [ 862.588199] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.589806] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.589806] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.591740] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 862.591740] env[61006]: value = "task-1337099" [ 862.591740] env[61006]: _type = "Task" [ 862.591740] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.602322] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337095, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.613070] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337098, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.613316] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.676460] env[61006]: DEBUG nova.policy [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8865ab804d34ccba8909c60cdb51adb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f60c05599bb7457f9bd7a3d11daf9ab3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.057793] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337096, 'name': ReconfigVM_Task, 'duration_secs': 0.709067} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.059069] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.059206] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8eff2451-928e-46be-b029-4b8fd97705f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.067639] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 863.067639] env[61006]: value = "task-1337100" [ 863.067639] env[61006]: _type = "Task" [ 863.067639] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.072382] env[61006]: DEBUG nova.objects.instance [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'pci_requests' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.089477] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337100, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.090873] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.105540] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337095, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.122498] env[61006]: DEBUG oslo_vmware.api [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337098, 'name': PowerOnVM_Task, 'duration_secs': 0.531663} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.126289] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.126607] env[61006]: INFO nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Took 12.02 seconds to spawn the instance on the hypervisor. [ 863.127168] env[61006]: DEBUG nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.127863] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.131250] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d900a724-e7b8-4025-989a-1685e14de308 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.291772] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Successfully created port: 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.375619] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d634df-95b2-4560-9f36-b3849fafc683 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.384159] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f64c2c6-eba8-482b-87b2-3ab9cab21241 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.416229] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6899ae9d-d533-4df1-a8d9-81d48646ec48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.425399] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dcd35d-6209-4825-90a9-b3a441c7df56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.442201] env[61006]: DEBUG nova.compute.provider_tree [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.576279] env[61006]: DEBUG nova.objects.base [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 863.576279] env[61006]: DEBUG nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.583172] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337100, 'name': Rename_Task, 'duration_secs': 0.159362} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.588232] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.588903] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55a7377b-8621-4015-9c0f-ab78b17b0fc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.608921] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337095, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.987382} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.614020] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] d174a8ec-867e-4fea-b878-2a9af1476949/d174a8ec-867e-4fea-b878-2a9af1476949.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.614020] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.614020] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 863.614020] env[61006]: value = "task-1337101" [ 863.614020] env[61006]: _type = "Task" [ 863.614020] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.615449] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0429bf72-da7b-412b-b174-d3e0f7ba1259 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.630783] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337101, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.635310] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.635625] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 863.635625] env[61006]: value = "task-1337102" [ 863.635625] env[61006]: _type = "Task" [ 863.635625] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.649695] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.653329] env[61006]: INFO nova.compute.manager [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Took 30.87 seconds to build instance. [ 863.725060] env[61006]: DEBUG nova.policy [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.945887] env[61006]: DEBUG nova.scheduler.client.report [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.112286] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.127800] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337099, 'name': PowerOffVM_Task, 'duration_secs': 1.460168} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.133151] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.133487] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337101, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.136269] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09025f5-e950-4b6b-bdee-6fb2794a4350 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.146909] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.147164] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.147323] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.147503] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.147680] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.147848] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.148239] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.148625] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.148625] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.148801] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.148988] env[61006]: DEBUG nova.virt.hardware [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.149790] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99459553-3caa-4faa-a435-03c009f46817 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.168269] env[61006]: DEBUG oslo_concurrency.lockutils [None req-43b87f9e-5cdf-4ec0-899b-35e9e9b25caf tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.675s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.169566] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c601bbac-9d98-4ba1-b2e6-ab5506a13e20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.178369] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077719} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.179242] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b76858-9f6d-4543-9100-467ee3cba64c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.183454] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.186315] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d48dae7-db8f-41b0-a35f-bd43d7f46445 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.218808] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] d174a8ec-867e-4fea-b878-2a9af1476949/d174a8ec-867e-4fea-b878-2a9af1476949.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.221029] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c951f0f-6cde-45a8-939c-d20267e04630 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.245161] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.245591] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee3b6c48-413e-4785-8ae0-d476176c3b30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.248458] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 864.248458] env[61006]: value = "task-1337103" [ 864.248458] env[61006]: _type = "Task" [ 864.248458] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.253761] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 864.253761] env[61006]: value = "task-1337104" [ 864.253761] env[61006]: _type = "Task" [ 864.253761] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.261551] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337103, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.268435] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 864.268435] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.268646] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.268842] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.269568] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.269831] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19206c6b-c398-41f6-8999-aafdb2674bdd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.284731] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.284731] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.284908] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cc0ba46-65c7-4016-aaec-27c0aa38598d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.290270] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 864.290270] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52db060c-4f45-89a5-4fe6-01f3c519dc05" [ 864.290270] env[61006]: _type = "Task" [ 864.290270] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.298659] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52db060c-4f45-89a5-4fe6-01f3c519dc05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.452673] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.452673] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.456988] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.312s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.456988] env[61006]: DEBUG nova.objects.instance [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lazy-loading 'resources' on Instance uuid 72c3e37b-2eac-41d4-8308-0a6466c2dd24 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.630090] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337101, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.762054] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.802969] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52db060c-4f45-89a5-4fe6-01f3c519dc05, 'name': SearchDatastore_Task, 'duration_secs': 0.046289} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.803956] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5e4288-42c6-4ec8-9095-b10bfea959d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.810568] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 864.810568] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d77ee-2924-34cd-de1b-28631544d98a" [ 864.810568] env[61006]: _type = "Task" [ 864.810568] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.824971] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d77ee-2924-34cd-de1b-28631544d98a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.961405] env[61006]: DEBUG nova.compute.utils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.963265] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.963820] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.037304] env[61006]: DEBUG nova.policy [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.134462] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337101, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.274492] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337103, 'name': ReconfigVM_Task, 'duration_secs': 0.813006} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.275358] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfigured VM instance instance-0000004c to attach disk [datastore1] d174a8ec-867e-4fea-b878-2a9af1476949/d174a8ec-867e-4fea-b878-2a9af1476949.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.276248] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c1b1397-9225-4056-83b1-0e4ff793eb71 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.286127] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 865.286127] env[61006]: value = "task-1337105" [ 865.286127] env[61006]: _type = "Task" [ 865.286127] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.287734] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b89066-ec26-4264-be44-9cd4418595c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.297352] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Successfully updated port: 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.307298] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4d91e2-cc01-4d91-a4f0-384be56e97e2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.311892] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337105, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.326416] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527d77ee-2924-34cd-de1b-28631544d98a, 'name': SearchDatastore_Task, 'duration_secs': 0.011525} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.358022] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.358022] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. {{(pid=61006) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 865.358022] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d777b5a0-daa8-4a09-b7de-744ad7462b69 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.359856] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e527f0-7d18-49a3-bc1b-35ad4205102e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.369995] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045219f6-76d6-4d13-bfc0-f409aaa395b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.374321] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 865.374321] env[61006]: value = "task-1337106" [ 865.374321] env[61006]: _type = "Task" [ 865.374321] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.385236] env[61006]: DEBUG nova.compute.provider_tree [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.393390] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.468168] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.485676] env[61006]: DEBUG nova.compute.manager [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.485676] env[61006]: DEBUG oslo_concurrency.lockutils [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.485676] env[61006]: DEBUG oslo_concurrency.lockutils [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.485832] env[61006]: DEBUG oslo_concurrency.lockutils [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.485964] env[61006]: DEBUG nova.compute.manager [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] No waiting events found dispatching network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.486319] env[61006]: WARNING nova.compute.manager [req-2f0f8015-da2c-49fc-aabb-cefa89e42018 req-e15fbe86-cc7a-494e-8631-47faa68ac619 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received unexpected event network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba for instance with vm_state building and task_state spawning. [ 865.634047] env[61006]: DEBUG oslo_vmware.api [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337101, 'name': PowerOnVM_Task, 'duration_secs': 1.618861} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.634360] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.634546] env[61006]: INFO nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Took 12.11 seconds to spawn the instance on the hypervisor. [ 865.634740] env[61006]: DEBUG nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.636147] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cbb426-cd70-4bce-8c7b-dde2d9d2f7a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.781876] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Successfully created port: 5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.800641] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.801156] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.801156] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.805457] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337105, 'name': Rename_Task, 'duration_secs': 0.212844} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.805457] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.805457] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92f3c5be-0063-470a-beda-37024843d4bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.813118] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 865.813118] env[61006]: value = "task-1337107" [ 865.813118] env[61006]: _type = "Task" [ 865.813118] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.822153] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.884722] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513302} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.885087] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. [ 865.885918] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bd9359-b8ba-4c79-a6d7-93abb2727031 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.889408] env[61006]: DEBUG nova.scheduler.client.report [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.896428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.896428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.896428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.896428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.896428] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.921879] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.923435] env[61006]: INFO nova.compute.manager [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Terminating instance [ 865.925012] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dec86594-e057-4b74-8286-4078e2e03089 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.943594] env[61006]: DEBUG nova.compute.manager [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.943881] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.944796] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f468354-729a-4e01-b0fb-cf2c755c35ac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.953864] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.955304] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7a0bdaa-8f98-479c-9980-4fc5cf9b425c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.957013] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 865.957013] env[61006]: value = "task-1337108" [ 865.957013] env[61006]: _type = "Task" [ 865.957013] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.963318] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 865.963318] env[61006]: value = "task-1337109" [ 865.963318] env[61006]: _type = "Task" [ 865.963318] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.970449] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.980302] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337109, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.064630] env[61006]: DEBUG nova.compute.manager [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-plugged-385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.064854] env[61006]: DEBUG oslo_concurrency.lockutils [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.065152] env[61006]: DEBUG oslo_concurrency.lockutils [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.065401] env[61006]: DEBUG oslo_concurrency.lockutils [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.065620] env[61006]: DEBUG nova.compute.manager [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] No waiting events found dispatching network-vif-plugged-385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.065834] env[61006]: WARNING nova.compute.manager [req-7618dc67-1d6b-4639-ab00-22663a7ddde6 req-93a6813b-e3d9-4e73-a410-94465bd22a4d service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received unexpected event network-vif-plugged-385eff2a-3efb-49b9-8063-c01a39357798 for instance with vm_state active and task_state None. [ 866.154732] env[61006]: INFO nova.compute.manager [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Took 32.83 seconds to build instance. [ 866.207748] env[61006]: DEBUG nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Successfully updated port: 385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.324214] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337107, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.359471] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.396226] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.399936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 21.173s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.399936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.399936] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 866.399936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.058s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.404769] env[61006]: INFO nova.compute.claims [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.408284] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7702164-dcc9-4ab1-a6c0-67ff78df63d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.420019] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f389349a-a0da-4191-84d6-8e23adc530d4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.438741] env[61006]: INFO nova.scheduler.client.report [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Deleted allocations for instance 72c3e37b-2eac-41d4-8308-0a6466c2dd24 [ 866.442771] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d08d87a-c522-437c-b17d-64f39486f50f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.455713] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c06c94-0556-41c6-be3c-60f07c999d29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.501079] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.503851] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180482MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 866.504037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.509293] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337108, 'name': ReconfigVM_Task, 'duration_secs': 0.380455} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.512271] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.516618] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fac3d1b-b966-4861-a0c2-c0ee2d072ab2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.519737] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337109, 'name': PowerOffVM_Task, 'duration_secs': 0.193481} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.520478] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.520548] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.521664] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0fc03ff-854b-4c6c-aac4-75db09fbbe58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.551026] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7fb7734-5532-4d99-a1d3-680316ba7dfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.568251] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 866.568251] env[61006]: value = "task-1337111" [ 866.568251] env[61006]: _type = "Task" [ 866.568251] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.574306] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.574306] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.574523] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.574642] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.574880] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.574970] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.575666] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.575666] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.575666] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.575817] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.575856] env[61006]: DEBUG nova.virt.hardware [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.576710] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c5dcd5-e7e4-4235-bce6-8c3f94aede91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.590678] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2529176-d952-4680-8575-ccd85454c79f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.595747] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337111, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.597612] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.597996] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.598316] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleting the datastore file [datastore1] 9c9fa347-bcfe-4009-af72-5f427e9d234a {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.599147] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62c2d8a1-9bec-4618-82bd-739ed9f460e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.614072] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for the task: (returnval){ [ 866.614072] env[61006]: value = "task-1337112" [ 866.614072] env[61006]: _type = "Task" [ 866.614072] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.622710] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.657142] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa28f547-1969-4786-9dd5-612e615f720f tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.302s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.700515] env[61006]: DEBUG nova.network.neutron [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.707149] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.707328] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.707507] env[61006]: DEBUG nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.825117] env[61006]: DEBUG oslo_vmware.api [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337107, 'name': PowerOnVM_Task, 'duration_secs': 0.880003} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.825478] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.825689] env[61006]: INFO nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Took 10.87 seconds to spawn the instance on the hypervisor. [ 866.825896] env[61006]: DEBUG nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.826798] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3415056c-3773-4fa8-b4ce-7e143dcb3c26 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.952699] env[61006]: DEBUG oslo_concurrency.lockutils [None req-feb87479-34b0-4f2c-86bd-b562f548141f tempest-VolumesAdminNegativeTest-392542767 tempest-VolumesAdminNegativeTest-392542767-project-member] Lock "72c3e37b-2eac-41d4-8308-0a6466c2dd24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.802s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.079428] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337111, 'name': ReconfigVM_Task, 'duration_secs': 0.375331} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.079746] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.080028] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a85c145d-bd77-482b-ac28-132f3ff8337e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.089552] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 867.089552] env[61006]: value = "task-1337113" [ 867.089552] env[61006]: _type = "Task" [ 867.089552] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.098356] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.124590] env[61006]: DEBUG oslo_vmware.api [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Task: {'id': task-1337112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377423} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.124890] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.125100] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.125913] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.125913] env[61006]: INFO nova.compute.manager [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Took 1.18 seconds to destroy the instance on the hypervisor. [ 867.125913] env[61006]: DEBUG oslo.service.loopingcall [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.125913] env[61006]: DEBUG nova.compute.manager [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.126105] env[61006]: DEBUG nova.network.neutron [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.203532] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.205382] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance network_info: |[{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.206350] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:76:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9230a732-200c-4084-8c6c-a5892e9a50ba', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.223370] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating folder: Project (f60c05599bb7457f9bd7a3d11daf9ab3). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.229326] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b531c4c7-319a-4764-8bb4-1d71817b5f52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.241169] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created folder: Project (f60c05599bb7457f9bd7a3d11daf9ab3) in parent group-v285275. [ 867.242105] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating folder: Instances. Parent ref: group-v285352. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.242431] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c94ba2aa-bd1c-4a27-9f76-518876cfde31 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.255447] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created folder: Instances in parent group-v285352. [ 867.256313] env[61006]: DEBUG oslo.service.loopingcall [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.256594] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.257062] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b763aa66-ee89-4ff1-af30-8e5ea71f331a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.279364] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.279364] env[61006]: value = "task-1337116" [ 867.279364] env[61006]: _type = "Task" [ 867.279364] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.286029] env[61006]: WARNING nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] 98dcba29-9f32-40eb-b526-46e091c2a330 already exists in list: networks containing: ['98dcba29-9f32-40eb-b526-46e091c2a330']. ignoring it [ 867.286336] env[61006]: WARNING nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] 98dcba29-9f32-40eb-b526-46e091c2a330 already exists in list: networks containing: ['98dcba29-9f32-40eb-b526-46e091c2a330']. ignoring it [ 867.293628] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337116, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.358742] env[61006]: INFO nova.compute.manager [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Took 31.62 seconds to build instance. [ 867.604533] env[61006]: DEBUG oslo_vmware.api [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337113, 'name': PowerOnVM_Task, 'duration_secs': 0.445295} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.607416] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.610683] env[61006]: DEBUG nova.compute.manager [None req-b8e487f5-2c3e-4d4d-b91d-89d16eaee8c6 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.614477] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c73dcd-bbc1-49ba-be0c-95f2d20ca4a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.790626] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337116, 'name': CreateVM_Task, 'duration_secs': 0.484544} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.791313] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.791871] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3b3727-a8f3-453c-9e66-83790d5794f1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.795064] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.795257] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.795525] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.795770] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e27e196f-8a85-43f0-9f9f-2eff7a283ab2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.804276] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f62e953-2fe8-434a-806f-41eb31ed0899 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.807935] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 867.807935] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525e3b26-6d84-656a-3d06-951c6b43fb64" [ 867.807935] env[61006]: _type = "Task" [ 867.807935] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.839028] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de293e13-f07d-4e82-85ca-fda4780545a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.845638] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525e3b26-6d84-656a-3d06-951c6b43fb64, 'name': SearchDatastore_Task, 'duration_secs': 0.011958} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.846299] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.846590] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.846833] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.847066] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.847163] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.847420] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70f7e3ea-59c6-4c9a-a786-d1ca692ccdc6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.852635] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef142c0a-9d1b-4572-ad3f-891bc94d40e8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.858146] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.858337] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.859536] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f72f5b4-3611-4f3c-b0fb-54ddb8674bcc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.870826] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7dcf2a76-e02c-436d-9226-c146faca4462 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.475s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.871257] env[61006]: DEBUG nova.compute.provider_tree [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.878616] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 867.878616] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2fe4d-58e8-6f4a-0e72-cb186eb3c3ac" [ 867.878616] env[61006]: _type = "Task" [ 867.878616] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.888384] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2fe4d-58e8-6f4a-0e72-cb186eb3c3ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.072691] env[61006]: DEBUG nova.network.neutron [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.116980] env[61006]: DEBUG nova.compute.manager [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.117857] env[61006]: DEBUG nova.compute.manager [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing instance network info cache due to event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.118105] env[61006]: DEBUG oslo_concurrency.lockutils [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.120503] env[61006]: DEBUG oslo_concurrency.lockutils [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.120503] env[61006]: DEBUG nova.network.neutron [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.146501] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Successfully updated port: 5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.173086] env[61006]: DEBUG nova.network.neutron [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "address": "fa:16:3e:2b:cd:e9", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0df9528c-28", "ovs_interfaceid": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "385eff2a-3efb-49b9-8063-c01a39357798", "address": "fa:16:3e:35:6b:93", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap385eff2a-3e", "ovs_interfaceid": "385eff2a-3efb-49b9-8063-c01a39357798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.376884] env[61006]: DEBUG nova.scheduler.client.report [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.394136] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c2fe4d-58e8-6f4a-0e72-cb186eb3c3ac, 'name': SearchDatastore_Task, 'duration_secs': 0.013803} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.397633] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f05092e0-817a-47cb-8d8e-86403d4ab7fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.405831] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 868.405831] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bbce26-76a6-db3c-ea48-01deb0379cee" [ 868.405831] env[61006]: _type = "Task" [ 868.405831] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.417479] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bbce26-76a6-db3c-ea48-01deb0379cee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.577179] env[61006]: INFO nova.compute.manager [-] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Took 1.45 seconds to deallocate network for instance. [ 868.653942] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.653942] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.653942] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.679137] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.679682] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.680337] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.681458] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67657af-70fc-411b-9b18-09dd94c67b5c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.702244] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.702244] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.702244] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.702610] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.702756] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.702969] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.703276] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.703511] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.703736] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.703976] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.704243] env[61006]: DEBUG nova.virt.hardware [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.711165] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfiguring VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 868.711952] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2915f8ae-0451-4de5-be49-4927a7585d32 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.729085] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-changed-385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.732404] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing instance network info cache due to event network-changed-385eff2a-3efb-49b9-8063-c01a39357798. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.732654] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.732794] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.733123] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Refreshing network info cache for port 385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.742311] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 868.742311] env[61006]: value = "task-1337117" [ 868.742311] env[61006]: _type = "Task" [ 868.742311] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.754411] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337117, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.888756] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.889621] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.897393] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.528s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.897393] env[61006]: DEBUG nova.objects.instance [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lazy-loading 'resources' on Instance uuid 28b83ea7-5433-472d-9e47-f73a4f2fb389 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.921888] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bbce26-76a6-db3c-ea48-01deb0379cee, 'name': SearchDatastore_Task, 'duration_secs': 0.030526} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.922182] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.922437] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.922970] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-750f2d12-d8ff-4bba-bb5c-58bfdeeb0910 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.933129] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 868.933129] env[61006]: value = "task-1337118" [ 868.933129] env[61006]: _type = "Task" [ 868.933129] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.942804] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.993783] env[61006]: DEBUG nova.network.neutron [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updated VIF entry in instance network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.994242] env[61006]: DEBUG nova.network.neutron [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.085567] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.186049] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.253916] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337117, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.368023] env[61006]: DEBUG nova.network.neutron [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Updating instance_info_cache with network_info: [{"id": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "address": "fa:16:3e:7d:f6:e0", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5712d7c4-e7", "ovs_interfaceid": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.405920] env[61006]: DEBUG nova.compute.utils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.412330] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.412853] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.455800] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337118, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.486461] env[61006]: DEBUG nova.policy [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82aea0ad3a245af99868fc53a0b04f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd26132a67fc1412c938a36094a34a433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.498075] env[61006]: DEBUG oslo_concurrency.lockutils [req-12afd604-f1b1-436a-87b7-4e86af484588 req-f39a2d3d-400e-435e-9c03-bf798e696073 service nova] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.673092] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updated VIF entry in instance network info cache for port 385eff2a-3efb-49b9-8063-c01a39357798. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.673349] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "address": "fa:16:3e:2b:cd:e9", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0df9528c-28", "ovs_interfaceid": "0df9528c-287c-45f7-8d8d-b43dc00c0b24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "385eff2a-3efb-49b9-8063-c01a39357798", "address": "fa:16:3e:35:6b:93", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap385eff2a-3e", "ovs_interfaceid": "385eff2a-3efb-49b9-8063-c01a39357798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.756196] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337117, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.789882] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293d3227-41b2-429f-a479-022ad83b55c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.799929] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8fdb2a-4d16-4360-98b8-715cd06428a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.836277] env[61006]: INFO nova.compute.manager [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Unrescuing [ 869.836597] env[61006]: DEBUG oslo_concurrency.lockutils [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.836902] env[61006]: DEBUG oslo_concurrency.lockutils [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.836996] env[61006]: DEBUG nova.network.neutron [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.838733] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cca03a-67af-4fe1-84d6-74420c768056 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.851098] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b45177b-c925-4d5d-afc1-6fa8719ed456 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.868963] env[61006]: DEBUG nova.compute.provider_tree [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.871861] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.871861] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Instance network_info: |[{"id": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "address": "fa:16:3e:7d:f6:e0", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5712d7c4-e7", "ovs_interfaceid": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.875314] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:f6:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37333dc2-982e-45e9-9dda-0c18417d7fa6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5712d7c4-e719-4ef4-9d1c-b827b35dff58', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.883515] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Creating folder: Project (8afbe151c21743bfb40dc12ba384db28). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.883820] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.884152] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.888063] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acc3f783-c4c8-4098-8590-bad4403087ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.900662] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Created folder: Project (8afbe151c21743bfb40dc12ba384db28) in parent group-v285275. [ 869.900999] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Creating folder: Instances. Parent ref: group-v285355. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.901528] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40c07da1-ff4c-4a7b-bc0d-55a852c41de7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.913084] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.916520] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Created folder: Instances in parent group-v285355. [ 869.916786] env[61006]: DEBUG oslo.service.loopingcall [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.917212] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.917456] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5930ce4e-1c24-43ac-a862-bb9d8a3bbff8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.944353] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730223} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.944637] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.944862] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.945150] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-254285c1-1458-4d64-bb8a-c46e23e885b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.950920] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.950920] env[61006]: value = "task-1337121" [ 869.950920] env[61006]: _type = "Task" [ 869.950920] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.954547] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 869.954547] env[61006]: value = "task-1337122" [ 869.954547] env[61006]: _type = "Task" [ 869.954547] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.961207] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337121, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.968632] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337122, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.016285] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Successfully created port: 8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.178410] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.178686] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Received event network-vif-deleted-953380b2-e816-4672-b9cd-2706604e550b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.178868] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Received event network-vif-plugged-5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.179072] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Acquiring lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.179282] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.179595] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.179762] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] No waiting events found dispatching network-vif-plugged-5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.179970] env[61006]: WARNING nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Received unexpected event network-vif-plugged-5712d7c4-e719-4ef4-9d1c-b827b35dff58 for instance with vm_state building and task_state spawning. [ 870.180267] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Received event network-changed-5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.180352] env[61006]: DEBUG nova.compute.manager [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Refreshing instance network info cache due to event network-changed-5712d7c4-e719-4ef4-9d1c-b827b35dff58. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.180545] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Acquiring lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.180686] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Acquired lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.180845] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Refreshing network info cache for port 5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.259433] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337117, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.377362] env[61006]: DEBUG nova.scheduler.client.report [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.389696] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.462711] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337121, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.468159] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337122, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085388} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.470760] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.471586] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a24263e-ed95-4fe6-ac77-31c6bdda6b2e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.496891] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.498076] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9104537-406e-4ef0-9090-f5e8ce6000b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.522773] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 870.522773] env[61006]: value = "task-1337123" [ 870.522773] env[61006]: _type = "Task" [ 870.522773] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.532198] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.594396] env[61006]: DEBUG nova.network.neutron [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [{"id": "ba96f533-658c-49f4-a9a4-48e129d5c197", "address": "fa:16:3e:bb:35:70", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba96f533-65", "ovs_interfaceid": "ba96f533-658c-49f4-a9a4-48e129d5c197", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.759062] env[61006]: DEBUG oslo_vmware.api [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337117, 'name': ReconfigVM_Task, 'duration_secs': 1.619133} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.759062] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.759062] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfigured VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 870.884350] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.886968] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.813s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.887272] env[61006]: DEBUG nova.objects.instance [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lazy-loading 'resources' on Instance uuid 6e99894e-81b6-4a07-9ec7-caa16272b3ba {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.913831] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.915419] env[61006]: INFO nova.scheduler.client.report [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Deleted allocations for instance 28b83ea7-5433-472d-9e47-f73a4f2fb389 [ 870.928856] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Updated VIF entry in instance network info cache for port 5712d7c4-e719-4ef4-9d1c-b827b35dff58. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.929234] env[61006]: DEBUG nova.network.neutron [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Updating instance_info_cache with network_info: [{"id": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "address": "fa:16:3e:7d:f6:e0", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5712d7c4-e7", "ovs_interfaceid": "5712d7c4-e719-4ef4-9d1c-b827b35dff58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.934024] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.970621] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337121, 'name': CreateVM_Task, 'duration_secs': 0.99243} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.972899] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8d2a59dff6373ad374128f48363d1f8c',container_format='bare',created_at=2024-10-22T16:13:37Z,direct_url=,disk_format='vmdk',id=06d8ddf9-1620-4813-97cd-7b51789db884,min_disk=1,min_ram=0,name='tempest-test-snap-1969237078',owner='d26132a67fc1412c938a36094a34a433',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-22T16:13:51Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.973651] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.974778] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.974975] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.975174] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.975507] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.975507] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.975793] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.975854] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.976025] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.978601] env[61006]: DEBUG nova.virt.hardware [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.978601] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.979678] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3065ee-b7bc-4cb7-b06d-aa93fa21c90d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.984075] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.984971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.985368] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.986275] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da8eace4-4a85-458b-b964-5741f552aa4a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.993013] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 870.993013] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529abf20-dbac-59a8-4143-bdf4bb621c0e" [ 870.993013] env[61006]: _type = "Task" [ 870.993013] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.999592] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d76ae8-480e-4222-97eb-4c2c852e6fc8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.010964] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529abf20-dbac-59a8-4143-bdf4bb621c0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.033300] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337123, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.043594] env[61006]: DEBUG nova.compute.manager [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Stashing vm_state: active {{(pid=61006) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 871.097747] env[61006]: DEBUG oslo_concurrency.lockutils [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-0111452e-1b4f-499c-932d-f31364d1a14c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.098780] env[61006]: DEBUG nova.objects.instance [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'flavor' on Instance uuid 0111452e-1b4f-499c-932d-f31364d1a14c {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.264118] env[61006]: DEBUG oslo_concurrency.lockutils [None req-73600c71-0438-4eff-8917-010b438100ac tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-385eff2a-3efb-49b9-8063-c01a39357798" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.818s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.428501] env[61006]: DEBUG oslo_concurrency.lockutils [None req-58799360-27ce-4e03-bbc1-afc84d45eee8 tempest-ServerPasswordTestJSON-1191986486 tempest-ServerPasswordTestJSON-1191986486-project-member] Lock "28b83ea7-5433-472d-9e47-f73a4f2fb389" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.052s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.436312] env[61006]: DEBUG oslo_concurrency.lockutils [req-78070aef-e5a4-45b7-ba34-ae08ef2ebd9c req-370ceb85-bf1c-4098-87ff-3fe5f6b3c9a6 service nova] Releasing lock "refresh_cache-4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.516338] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529abf20-dbac-59a8-4143-bdf4bb621c0e, 'name': SearchDatastore_Task, 'duration_secs': 0.039293} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.516338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.517084] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.517084] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.517191] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.517369] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.517802] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-995628c2-f69f-42f3-aae8-eff1c4a21541 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.531423] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.531641] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.532941] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12a441dd-6381-4998-83ef-0276d9d0cb2d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.538687] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337123, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.543583] env[61006]: DEBUG nova.compute.manager [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Received event network-changed-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.543583] env[61006]: DEBUG nova.compute.manager [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Refreshing instance network info cache due to event network-changed-0e26f4ec-25ca-4a63-8c6a-b6767eda7456. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.543583] env[61006]: DEBUG oslo_concurrency.lockutils [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] Acquiring lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.543700] env[61006]: DEBUG oslo_concurrency.lockutils [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] Acquired lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.543854] env[61006]: DEBUG nova.network.neutron [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Refreshing network info cache for port 0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.549624] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 871.549624] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ac6665-6a96-2979-0d03-2d8310088eaa" [ 871.549624] env[61006]: _type = "Task" [ 871.549624] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.562330] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ac6665-6a96-2979-0d03-2d8310088eaa, 'name': SearchDatastore_Task, 'duration_secs': 0.014286} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.563714] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-398b64e4-ec7a-403a-9c80-c85ffe26213e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.571912] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 871.571912] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525e549b-6df5-8d8a-99d0-3f2eba4b8cff" [ 871.571912] env[61006]: _type = "Task" [ 871.571912] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.575606] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.581910] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525e549b-6df5-8d8a-99d0-3f2eba4b8cff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.608364] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b92c21-cd89-4236-a7b8-cb7086731aea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.633682] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.634373] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd259864-3cdd-4807-a357-c5f85b5b5a03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.643421] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 871.643421] env[61006]: value = "task-1337124" [ 871.643421] env[61006]: _type = "Task" [ 871.643421] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.653358] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.752743] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8166c58b-d68e-4f3a-8c65-2fcda696a5af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.766165] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621b4487-66ca-4eeb-ba19-c7a6d06c95e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.799143] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65450a22-9c0c-481d-867c-38e7b56c5da5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.808830] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9152ad3-e755-4ce9-be5e-cef0157d127e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.824615] env[61006]: DEBUG nova.compute.provider_tree [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.035181] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337123, 'name': ReconfigVM_Task, 'duration_secs': 1.050867} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.035486] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.036068] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd0a14dd-64eb-43b5-8b0e-b0e82db0f0bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.043072] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 872.043072] env[61006]: value = "task-1337125" [ 872.043072] env[61006]: _type = "Task" [ 872.043072] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.055279] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337125, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.082994] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525e549b-6df5-8d8a-99d0-3f2eba4b8cff, 'name': SearchDatastore_Task, 'duration_secs': 0.011291} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.083310] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.083577] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e/4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.083887] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c62b7eca-3be7-499f-9fc9-f5c69b0a06ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.092393] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 872.092393] env[61006]: value = "task-1337126" [ 872.092393] env[61006]: _type = "Task" [ 872.092393] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.101424] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337126, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.155621] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337124, 'name': PowerOffVM_Task, 'duration_secs': 0.208305} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.155980] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.162472] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 872.162646] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-597614ff-115e-4e48-aef7-5a91eb0cda5b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.184936] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 872.184936] env[61006]: value = "task-1337127" [ 872.184936] env[61006]: _type = "Task" [ 872.184936] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.197233] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337127, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.210308] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Successfully updated port: 8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.328832] env[61006]: DEBUG nova.scheduler.client.report [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.555209] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337125, 'name': Rename_Task, 'duration_secs': 0.200239} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.555575] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.556024] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15241da2-485d-46c1-86a0-3d35df6c242e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.565558] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 872.565558] env[61006]: value = "task-1337128" [ 872.565558] env[61006]: _type = "Task" [ 872.565558] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.576134] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337128, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.604550] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337126, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.695995] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337127, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.713501] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.713728] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.713906] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.726790] env[61006]: DEBUG nova.network.neutron [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updated VIF entry in instance network info cache for port 0e26f4ec-25ca-4a63-8c6a-b6767eda7456. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.727261] env[61006]: DEBUG nova.network.neutron [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating instance_info_cache with network_info: [{"id": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "address": "fa:16:3e:c0:e7:6d", "network": {"id": "3d20b60a-ee54-49f6-80e6-835258e52eac", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-743242808-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ad8b37b118c4c8a8fde488ffdc44621", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e26f4ec-25", "ovs_interfaceid": "0e26f4ec-25ca-4a63-8c6a-b6767eda7456", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.836494] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.949s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.838901] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.097s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.841561] env[61006]: INFO nova.compute.claims [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.878868] env[61006]: INFO nova.scheduler.client.report [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Deleted allocations for instance 6e99894e-81b6-4a07-9ec7-caa16272b3ba [ 873.082760] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337128, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.110141] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337126, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590341} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.110894] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e/4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.113028] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.113028] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fec96060-890f-4027-8666-93b31dc45267 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.128394] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 873.128394] env[61006]: value = "task-1337129" [ 873.128394] env[61006]: _type = "Task" [ 873.128394] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.137109] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.183633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "42b92d52-d1f0-48ff-94b6-6164b387456c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.183843] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.184070] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.184259] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.184434] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.192698] env[61006]: INFO nova.compute.manager [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Terminating instance [ 873.196527] env[61006]: DEBUG nova.compute.manager [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.196722] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.201573] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c197b6ac-da4e-48af-b4e5-b4e4ba83e268 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.207999] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337127, 'name': ReconfigVM_Task, 'duration_secs': 0.607289} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.207999] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 873.207999] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.207999] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c675aee2-1588-47c8-bb2e-3945aca8cb88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.211278] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.211668] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93af2e34-5dd8-491b-b951-10aa5a295b8c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.214794] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 873.214794] env[61006]: value = "task-1337130" [ 873.214794] env[61006]: _type = "Task" [ 873.214794] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.221664] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 873.221664] env[61006]: value = "task-1337131" [ 873.221664] env[61006]: _type = "Task" [ 873.221664] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.232825] env[61006]: DEBUG oslo_concurrency.lockutils [req-0b249f3c-5f9b-4170-8416-0fcb64d2d118 req-edff5339-1ba2-4b43-9f4e-1bc351a27d41 service nova] Releasing lock "refresh_cache-d174a8ec-867e-4fea-b878-2a9af1476949" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.232825] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.239844] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1337131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.268414] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.394497] env[61006]: DEBUG oslo_concurrency.lockutils [None req-40f6de54-4eac-4840-8a8e-3e724ba6ad2b tempest-ServersV294TestFqdnHostnames-886876509 tempest-ServersV294TestFqdnHostnames-886876509-project-member] Lock "6e99894e-81b6-4a07-9ec7-caa16272b3ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.638s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.528791] env[61006]: DEBUG nova.network.neutron [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Updating instance_info_cache with network_info: [{"id": "8cc8c41b-62f9-4046-a1c9-764067e78608", "address": "fa:16:3e:86:6f:1f", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc8c41b-62", "ovs_interfaceid": "8cc8c41b-62f9-4046-a1c9-764067e78608", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.580893] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337128, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.614897] env[61006]: DEBUG nova.compute.manager [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Received event network-vif-plugged-8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.614897] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Acquiring lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.615291] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.616964] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.616964] env[61006]: DEBUG nova.compute.manager [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] No waiting events found dispatching network-vif-plugged-8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.616964] env[61006]: WARNING nova.compute.manager [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Received unexpected event network-vif-plugged-8cc8c41b-62f9-4046-a1c9-764067e78608 for instance with vm_state building and task_state spawning. [ 873.616964] env[61006]: DEBUG nova.compute.manager [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Received event network-changed-8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.616964] env[61006]: DEBUG nova.compute.manager [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Refreshing instance network info cache due to event network-changed-8cc8c41b-62f9-4046-a1c9-764067e78608. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.616964] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Acquiring lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.635335] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087393} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.635627] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.636462] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85827b64-7125-41b0-9bcb-3c12203d96ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.669247] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e/4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.670063] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d54b5057-5e61-4491-b490-0413232e9b6a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.695059] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 873.695059] env[61006]: value = "task-1337132" [ 873.695059] env[61006]: _type = "Task" [ 873.695059] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.706582] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337132, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.730024] env[61006]: DEBUG oslo_vmware.api [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337130, 'name': PowerOnVM_Task, 'duration_secs': 0.451896} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.731847] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.732340] env[61006]: DEBUG nova.compute.manager [None req-009088cc-85d7-4eea-b577-388764d5c08c tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.733254] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59e667c-1b14-47d2-bfc7-e4da2f1b48f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.745023] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1337131, 'name': PowerOffVM_Task, 'duration_secs': 0.387877} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.745023] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.745023] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.747721] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9aac6b7-b81c-4e61-a48c-ca1dc17f7265 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.824246] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.824598] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.824848] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Deleting the datastore file [datastore2] 42b92d52-d1f0-48ff-94b6-6164b387456c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.825680] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88776266-0ed5-41ce-b023-2d60dc945464 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.833999] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for the task: (returnval){ [ 873.833999] env[61006]: value = "task-1337134" [ 873.833999] env[61006]: _type = "Task" [ 873.833999] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.844701] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1337134, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.865045] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-0df9528c-287c-45f7-8d8d-b43dc00c0b24" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.866281] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-0df9528c-287c-45f7-8d8d-b43dc00c0b24" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.033858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.035693] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Instance network_info: |[{"id": "8cc8c41b-62f9-4046-a1c9-764067e78608", "address": "fa:16:3e:86:6f:1f", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc8c41b-62", "ovs_interfaceid": "8cc8c41b-62f9-4046-a1c9-764067e78608", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 874.035693] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Acquired lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.035693] env[61006]: DEBUG nova.network.neutron [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Refreshing network info cache for port 8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.036444] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:6f:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1931669-8959-4e86-a603-e206bcf2b47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cc8c41b-62f9-4046-a1c9-764067e78608', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.048081] env[61006]: DEBUG oslo.service.loopingcall [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.054936] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.056644] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d4107bd-7fdd-4ee9-ae6a-28ae930c1399 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.089193] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337128, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.090917] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.090917] env[61006]: value = "task-1337135" [ 874.090917] env[61006]: _type = "Task" [ 874.090917] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.102905] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337135, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.207019] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33886e52-f4dd-4204-8f02-bed8c074f365 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.213404] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337132, 'name': ReconfigVM_Task, 'duration_secs': 0.318981} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.215693] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e/4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.216412] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-adb1ade6-c381-4370-81b2-d59f4c276a16 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.222940] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad74ff2-0009-4dcf-8df6-23871febfc66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.230245] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 874.230245] env[61006]: value = "task-1337136" [ 874.230245] env[61006]: _type = "Task" [ 874.230245] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.271995] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d42d18-02de-4125-83d2-82db2ed0a97b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.277898] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337136, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.286552] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b34b6fc-c174-4731-ac91-daf1f9d8d0f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.304139] env[61006]: DEBUG nova.compute.provider_tree [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.348585] env[61006]: DEBUG oslo_vmware.api [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Task: {'id': task-1337134, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.376195} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.349064] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.349320] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.349625] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.349747] env[61006]: INFO nova.compute.manager [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 874.350153] env[61006]: DEBUG oslo.service.loopingcall [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.350310] env[61006]: DEBUG nova.compute.manager [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.350422] env[61006]: DEBUG nova.network.neutron [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.374420] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.374627] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.375566] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88e80a7-2440-4c6b-b95f-0076b442fe97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.399239] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30598d60-99dc-466e-bd9a-a1164ecf8478 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.424769] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfiguring VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 874.429499] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad14ca4a-abb8-40af-a5fa-c2ae38d2e27c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.449635] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 874.449635] env[61006]: value = "task-1337137" [ 874.449635] env[61006]: _type = "Task" [ 874.449635] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.462372] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.594537] env[61006]: DEBUG oslo_vmware.api [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337128, 'name': PowerOnVM_Task, 'duration_secs': 1.628144} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.598675] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.598941] env[61006]: INFO nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Took 10.49 seconds to spawn the instance on the hypervisor. [ 874.599329] env[61006]: DEBUG nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.600773] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39252a32-8c47-4a18-96e1-53bd8831c01f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.614605] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337135, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.620886] env[61006]: DEBUG nova.network.neutron [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Updated VIF entry in instance network info cache for port 8cc8c41b-62f9-4046-a1c9-764067e78608. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.621416] env[61006]: DEBUG nova.network.neutron [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Updating instance_info_cache with network_info: [{"id": "8cc8c41b-62f9-4046-a1c9-764067e78608", "address": "fa:16:3e:86:6f:1f", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cc8c41b-62", "ovs_interfaceid": "8cc8c41b-62f9-4046-a1c9-764067e78608", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.741279] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337136, 'name': Rename_Task, 'duration_secs': 0.157677} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.741742] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.742271] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c10f4884-0046-409b-8e18-008b14f4a1d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.753166] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 874.753166] env[61006]: value = "task-1337138" [ 874.753166] env[61006]: _type = "Task" [ 874.753166] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.764948] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.809389] env[61006]: DEBUG nova.scheduler.client.report [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.963519] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.105245] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337135, 'name': CreateVM_Task, 'duration_secs': 0.523354} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.105519] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.106094] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.106276] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.106644] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.106905] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2273cad2-703e-4ec3-9641-3d22236aeb03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.112228] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 875.112228] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b19fe-4295-4ce6-d895-8717da27cd24" [ 875.112228] env[61006]: _type = "Task" [ 875.112228] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.120778] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b19fe-4295-4ce6-d895-8717da27cd24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.128353] env[61006]: DEBUG oslo_concurrency.lockutils [req-b0b78a2f-0a3c-48e5-8ff9-232e16f132ce req-420efa51-6beb-46bf-8b25-7b301806214d service nova] Releasing lock "refresh_cache-2232c77d-9f4e-4981-9f55-614bba5d71db" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.130667] env[61006]: INFO nova.compute.manager [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Took 36.39 seconds to build instance. [ 875.262373] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337138, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.315413] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.316405] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.319269] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.733s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.319654] env[61006]: DEBUG nova.objects.instance [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lazy-loading 'resources' on Instance uuid 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.464938] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.629389] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.630151] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Processing image 06d8ddf9-1620-4813-97cd-7b51789db884 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.630601] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.630601] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.630815] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.631182] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1966312-5f18-4dd8-a4df-c407efda1359 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.634140] env[61006]: DEBUG oslo_concurrency.lockutils [None req-88419c1c-c66b-4f62-a860-6ffc619bdf4a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.148s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.644317] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.645025] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.645991] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-151d606f-04bf-43bb-88ff-3e0e19a90f27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.657218] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 875.657218] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52eec0b8-9355-41d2-9234-caa0927cd72a" [ 875.657218] env[61006]: _type = "Task" [ 875.657218] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.669861] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52eec0b8-9355-41d2-9234-caa0927cd72a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.763118] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337138, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.822890] env[61006]: DEBUG nova.compute.utils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.824531] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.824624] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.845409] env[61006]: DEBUG nova.compute.manager [req-5d04a78d-d940-4dc2-89d5-aabf55cfe3aa req-bbabc072-65eb-445f-ad0d-a8061ac87607 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Received event network-vif-deleted-0652e5ba-2f85-49a7-85aa-2f0bc1757b4a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.845409] env[61006]: INFO nova.compute.manager [req-5d04a78d-d940-4dc2-89d5-aabf55cfe3aa req-bbabc072-65eb-445f-ad0d-a8061ac87607 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Neutron deleted interface 0652e5ba-2f85-49a7-85aa-2f0bc1757b4a; detaching it from the instance and deleting it from the info cache [ 875.845409] env[61006]: DEBUG nova.network.neutron [req-5d04a78d-d940-4dc2-89d5-aabf55cfe3aa req-bbabc072-65eb-445f-ad0d-a8061ac87607 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.946011] env[61006]: DEBUG nova.network.neutron [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.963032] env[61006]: DEBUG nova.policy [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.968968] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.172291] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 876.172641] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Fetch image to [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c/OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 876.172765] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Downloading stream optimized image 06d8ddf9-1620-4813-97cd-7b51789db884 to [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c/OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c.vmdk on the data store datastore2 as vApp {{(pid=61006) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 876.172929] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Downloading image file data 06d8ddf9-1620-4813-97cd-7b51789db884 to the ESX as VM named 'OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c' {{(pid=61006) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 876.195629] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6fa250-db8b-4602-b86d-f30030f19cdc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.230520] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a11fd57-974f-490b-abba-9d3e31306a08 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.274724] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e08c75-2715-42b9-98f9-275c61da639d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.286799] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843cc6af-48ad-4e0e-8277-92f3a0fb3a43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.290625] env[61006]: DEBUG oslo_vmware.api [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337138, 'name': PowerOnVM_Task, 'duration_secs': 1.503687} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.290913] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.291133] env[61006]: INFO nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Took 9.79 seconds to spawn the instance on the hypervisor. [ 876.291314] env[61006]: DEBUG nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.292482] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8ef168-c6b8-4bc7-af31-6824a50350bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.304682] env[61006]: DEBUG nova.compute.provider_tree [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.306457] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 876.306457] env[61006]: value = "resgroup-9" [ 876.306457] env[61006]: _type = "ResourcePool" [ 876.306457] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 876.309349] env[61006]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-acf9042f-fb42-4c9d-9c19-75222f59e532 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.328941] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.335357] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lease: (returnval){ [ 876.335357] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 876.335357] env[61006]: _type = "HttpNfcLease" [ 876.335357] env[61006]: } obtained for vApp import into resource pool (val){ [ 876.335357] env[61006]: value = "resgroup-9" [ 876.335357] env[61006]: _type = "ResourcePool" [ 876.335357] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 876.335615] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the lease: (returnval){ [ 876.335615] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 876.335615] env[61006]: _type = "HttpNfcLease" [ 876.335615] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 876.344407] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 876.344407] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 876.344407] env[61006]: _type = "HttpNfcLease" [ 876.344407] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 876.352178] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27a234a0-20a5-4d70-9557-57d820da3dc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.363812] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638e7b0e-7d3a-41fd-ae36-aa91bb3fa802 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.403914] env[61006]: DEBUG nova.compute.manager [req-5d04a78d-d940-4dc2-89d5-aabf55cfe3aa req-bbabc072-65eb-445f-ad0d-a8061ac87607 service nova] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Detach interface failed, port_id=0652e5ba-2f85-49a7-85aa-2f0bc1757b4a, reason: Instance 42b92d52-d1f0-48ff-94b6-6164b387456c could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 876.448367] env[61006]: INFO nova.compute.manager [-] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Took 2.10 seconds to deallocate network for instance. [ 876.464649] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.541540] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Successfully created port: d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.824186] env[61006]: DEBUG nova.scheduler.client.report [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.850294] env[61006]: INFO nova.compute.manager [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Took 35.82 seconds to build instance. [ 876.858301] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 876.858301] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 876.858301] env[61006]: _type = "HttpNfcLease" [ 876.858301] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 876.962768] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.972569] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.341287] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.022s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.345198] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.810s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.347108] env[61006]: INFO nova.compute.claims [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.351149] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.353700] env[61006]: DEBUG oslo_concurrency.lockutils [None req-49325bd5-c170-4d74-b12a-0bcb425dadde tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.219s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.365095] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.365095] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 877.365095] env[61006]: _type = "HttpNfcLease" [ 877.365095] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.378776] env[61006]: INFO nova.scheduler.client.report [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Deleted allocations for instance 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a [ 877.399019] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.399263] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.399418] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.399656] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.399806] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.399950] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.400382] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.400627] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.400844] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.405018] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.405018] env[61006]: DEBUG nova.virt.hardware [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.405018] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a28d880-00ea-4ac0-a142-acc1e44eadcd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.411987] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ab5cd4-3491-4abf-bbd1-78ceb58f4846 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.465732] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.867315] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.867315] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 877.867315] env[61006]: _type = "HttpNfcLease" [ 877.867315] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 877.870943] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 877.870943] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52caab12-3d94-16e9-b02b-5f2de9ca7c85" [ 877.870943] env[61006]: _type = "HttpNfcLease" [ 877.870943] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 877.870943] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043acc75-999a-47d4-b0f1-d27039db4c03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.888051] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 877.888631] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 877.894376] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e16c1d1d-0f12-4c6c-b0b0-308ede020d17 tempest-ServersTestMultiNic-1931006950 tempest-ServersTestMultiNic-1931006950-project-member] Lock "1a7b5a77-b43e-4c96-ac18-36634a0e7d5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.605s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.978038] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e06d3f92-0b86-4c8b-9178-5ae9756761d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.980572] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.254486] env[61006]: DEBUG nova.compute.manager [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.254745] env[61006]: DEBUG nova.compute.manager [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing instance network info cache due to event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.255149] env[61006]: DEBUG oslo_concurrency.lockutils [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.255528] env[61006]: DEBUG oslo_concurrency.lockutils [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.255776] env[61006]: DEBUG nova.network.neutron [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.397382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.399152] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.469012] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.470174] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.470174] env[61006]: DEBUG nova.compute.manager [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.471536] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15f47ab-8591-4f39-bc26-fb00423ccef8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.483444] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.483928] env[61006]: DEBUG nova.compute.manager [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61006) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 878.486818] env[61006]: DEBUG nova.objects.instance [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lazy-loading 'flavor' on Instance uuid 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.768970] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb1f11a-a100-407f-b44b-8883c98b46bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.773940] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Successfully updated port: d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.781872] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9975c2aa-c44c-4c92-ba07-bf2b10fb06e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.843064] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a912e470-be1b-44fe-bb7f-1017f67de9e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.856976] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b9af1b-63b2-4e8b-bac1-c6daa8c8d06d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.878410] env[61006]: DEBUG nova.compute.provider_tree [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.900483] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 878.979124] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.998038] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.998038] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7f83ad2-dacc-44a4-b7ec-323722035c65 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.008802] env[61006]: DEBUG oslo_vmware.api [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 879.008802] env[61006]: value = "task-1337140" [ 879.008802] env[61006]: _type = "Task" [ 879.008802] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.028419] env[61006]: DEBUG oslo_vmware.api [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337140, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.076091] env[61006]: DEBUG nova.network.neutron [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updated VIF entry in instance network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.076654] env[61006]: DEBUG nova.network.neutron [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.147766] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 879.147949] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 879.148793] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9fc1e7-9c14-4aaa-b565-f2cad4a84502 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.156512] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 879.156751] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 879.157101] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-72feef84-99fc-44a2-baa7-271bd69273d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.277900] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.277992] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.278186] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.340078] env[61006]: DEBUG oslo_vmware.rw_handles [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5237275e-1af3-01bb-919b-5f0342e4ce69/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 879.340345] env[61006]: INFO nova.virt.vmwareapi.images [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Downloaded image file data 06d8ddf9-1620-4813-97cd-7b51789db884 [ 879.341616] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecafb13-7471-42d2-acf4-c90f2f832bb0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.360713] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a8ef257-d904-4e5b-8022-8a437134e856 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.383881] env[61006]: DEBUG nova.scheduler.client.report [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.407349] env[61006]: INFO nova.virt.vmwareapi.images [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] The imported VM was unregistered [ 879.408966] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 879.409212] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating directory with path [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.412092] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94214ca3-e500-4f9b-8ce9-900488d7e8cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.436274] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.449034] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created directory with path [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.449034] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c/OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c.vmdk to [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk. {{(pid=61006) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 879.449315] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9b59cb27-2d9a-4fcc-ad82-e6e45aaeca88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.457886] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 879.457886] env[61006]: value = "task-1337142" [ 879.457886] env[61006]: _type = "Task" [ 879.457886] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.466140] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.474555] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.519443] env[61006]: DEBUG oslo_vmware.api [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337140, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.582429] env[61006]: DEBUG oslo_concurrency.lockutils [req-2c8c9e9c-2b21-411f-a711-8002cfbd276d req-82592f5f-147f-4503-89e9-1795b2319ea9 service nova] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.602305] env[61006]: DEBUG nova.compute.manager [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.603245] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e14d50-28aa-42dc-96a8-643bbb23fc72 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.834843] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 879.889365] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.889934] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.892685] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.333s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.897502] env[61006]: INFO nova.compute.claims [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.969819] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.978790] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.025819] env[61006]: DEBUG oslo_vmware.api [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337140, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.119374] env[61006]: INFO nova.compute.manager [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] instance snapshotting [ 880.120197] env[61006]: DEBUG nova.objects.instance [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.146559] env[61006]: DEBUG nova.network.neutron [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Updating instance_info_cache with network_info: [{"id": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "address": "fa:16:3e:6c:34:61", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d7cece-12", "ovs_interfaceid": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.302121] env[61006]: DEBUG nova.compute.manager [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Received event network-vif-plugged-d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.302471] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Acquiring lock "d1e09ace-177d-430a-9e65-d01368815272-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.302787] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Lock "d1e09ace-177d-430a-9e65-d01368815272-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.303085] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Lock "d1e09ace-177d-430a-9e65-d01368815272-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.303338] env[61006]: DEBUG nova.compute.manager [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] No waiting events found dispatching network-vif-plugged-d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.304638] env[61006]: WARNING nova.compute.manager [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Received unexpected event network-vif-plugged-d6d7cece-1280-46c8-ad3b-68ec36e57ebf for instance with vm_state building and task_state spawning. [ 880.304856] env[61006]: DEBUG nova.compute.manager [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Received event network-changed-d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.305046] env[61006]: DEBUG nova.compute.manager [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Refreshing instance network info cache due to event network-changed-d6d7cece-1280-46c8-ad3b-68ec36e57ebf. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.305237] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Acquiring lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.410157] env[61006]: DEBUG nova.compute.utils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.413484] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.413484] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.469150] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.479557] env[61006]: DEBUG oslo_vmware.api [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337137, 'name': ReconfigVM_Task, 'duration_secs': 5.856292} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.479922] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.480251] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Reconfigured VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 880.531213] env[61006]: DEBUG oslo_vmware.api [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337140, 'name': PowerOffVM_Task, 'duration_secs': 1.221444} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.531640] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.531914] env[61006]: DEBUG nova.compute.manager [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.532971] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87437c24-9f07-4b96-93cd-0a31f0ec2fb7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.539798] env[61006]: DEBUG nova.policy [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42027a2dce454ca98e295dadca850eee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2b5a4bf20a40ee9b8688935de75dc1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.629209] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d756ce-3abe-4f5d-97bf-ca898bd541c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.657384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.657792] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Instance network_info: |[{"id": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "address": "fa:16:3e:6c:34:61", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d7cece-12", "ovs_interfaceid": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.658369] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Acquired lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.658718] env[61006]: DEBUG nova.network.neutron [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Refreshing network info cache for port d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.661085] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:34:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6d7cece-1280-46c8-ad3b-68ec36e57ebf', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.669576] env[61006]: DEBUG oslo.service.loopingcall [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.670494] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823578de-75b8-4e6e-b302-54a2cbe9aed1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.674043] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.674748] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-065d3652-b86b-49f9-ac8f-b53881c69097 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.699709] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.700478] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.705828] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.705828] env[61006]: value = "task-1337143" [ 880.705828] env[61006]: _type = "Task" [ 880.705828] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.719624] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337143, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.922821] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.981550] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.056635] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e8277573-ee00-4206-8f0f-58dad45c029e tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.587s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.204632] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 881.210521] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 881.212292] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-514312be-5820-4ac3-9591-87875994bba1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.233883] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337143, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.237629] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 881.237629] env[61006]: value = "task-1337144" [ 881.237629] env[61006]: _type = "Task" [ 881.237629] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.255424] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337144, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.260051] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Successfully created port: 0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.382147] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748b4b94-d6c9-4c6a-a476-66deb0a2eac4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.392841] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf80e4b-fc28-4333-bb97-a2dca8f652a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.442488] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7322e760-6eb7-4661-824c-67f0d6980817 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.453284] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b5c15e-c483-4aa3-bd72-85a3eb8e3651 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.474037] env[61006]: DEBUG nova.compute.provider_tree [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.481717] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.605725] env[61006]: DEBUG nova.network.neutron [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Updated VIF entry in instance network info cache for port d6d7cece-1280-46c8-ad3b-68ec36e57ebf. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.606327] env[61006]: DEBUG nova.network.neutron [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Updating instance_info_cache with network_info: [{"id": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "address": "fa:16:3e:6c:34:61", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d7cece-12", "ovs_interfaceid": "d6d7cece-1280-46c8-ad3b-68ec36e57ebf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.726444] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337143, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.746269] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.754953] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337144, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.941352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.941352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.941352] env[61006]: DEBUG nova.network.neutron [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.944487] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.985394] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.985394] env[61006]: DEBUG nova.scheduler.client.report [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.003031] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.003600] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.003600] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.003600] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.004320] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.004320] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.004885] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decb1814-0f96-4490-af57-b817b652b334 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.018876] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d99f475-1d99-4b95-8f3d-b9989c6b3605 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.109939] env[61006]: DEBUG oslo_concurrency.lockutils [req-8d013c1b-6b7d-4e2f-9fc4-ccdb4d9a2f5f req-81b3f25c-a912-4d66-a01d-48b03b4382be service nova] Releasing lock "refresh_cache-d1e09ace-177d-430a-9e65-d01368815272" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.223319] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337143, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.252821] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337144, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.335385] env[61006]: DEBUG nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-deleted-0df9528c-287c-45f7-8d8d-b43dc00c0b24 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.335758] env[61006]: INFO nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Neutron deleted interface 0df9528c-287c-45f7-8d8d-b43dc00c0b24; detaching it from the instance and deleting it from the info cache [ 882.341098] env[61006]: DEBUG nova.network.neutron [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "385eff2a-3efb-49b9-8063-c01a39357798", "address": "fa:16:3e:35:6b:93", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap385eff2a-3e", "ovs_interfaceid": "385eff2a-3efb-49b9-8063-c01a39357798", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.394524] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.394781] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.395159] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.395315] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.395516] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.398463] env[61006]: INFO nova.compute.manager [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Terminating instance [ 882.400420] env[61006]: DEBUG nova.compute.manager [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.400687] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.401631] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cead10-2e8f-4b9e-9bcc-4c3c8ea611ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.411403] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.411742] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f5cd3e2-9b78-4f0a-8699-988449c059cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.446763] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.447246] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.447593] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.447960] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.448090] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.453043] env[61006]: INFO nova.compute.manager [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Terminating instance [ 882.459246] env[61006]: DEBUG nova.compute.manager [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.459499] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.460519] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0e22d7-a00a-4047-a9b3-f25b6f42fc20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.478520] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.481183] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.481500] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f14e4063-af9b-4480-a7bb-cecb2d2c3882 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.490804] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 882.490804] env[61006]: value = "task-1337146" [ 882.490804] env[61006]: _type = "Task" [ 882.490804] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.491385] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.492058] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.510515] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.006s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.515997] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.516327] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.517055] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleting the datastore file [datastore2] 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.517368] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef1b2003-5b75-4959-b02a-4a03efd4e71f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.523125] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.534704] env[61006]: DEBUG oslo_vmware.api [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 882.534704] env[61006]: value = "task-1337147" [ 882.534704] env[61006]: _type = "Task" [ 882.534704] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.549013] env[61006]: DEBUG oslo_vmware.api [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.723568] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337143, 'name': CreateVM_Task, 'duration_secs': 1.898301} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.723785] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.724553] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.724757] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.725195] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.725484] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b08c61d8-0b9c-4673-aceb-6e41db41a475 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.730959] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 882.730959] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ec4e8-0fba-b90f-9c41-3ab15f802a1f" [ 882.730959] env[61006]: _type = "Task" [ 882.730959] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.743908] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ec4e8-0fba-b90f-9c41-3ab15f802a1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.752406] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337144, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.772608] env[61006]: INFO nova.network.neutron [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Port 0df9528c-287c-45f7-8d8d-b43dc00c0b24 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 882.772850] env[61006]: INFO nova.network.neutron [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Port 385eff2a-3efb-49b9-8063-c01a39357798 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 882.773245] env[61006]: DEBUG nova.network.neutron [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.844168] env[61006]: DEBUG oslo_concurrency.lockutils [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] Acquiring lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.970183] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337142, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.074463} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.970465] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c/OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c.vmdk to [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk. [ 882.970654] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Cleaning up location [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 882.970818] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_32e8af7b-a308-4a05-a324-541277e17e9c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.971093] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c96a3d22-ea03-4ec3-88f1-f23849e38acc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.977791] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 882.977791] env[61006]: value = "task-1337148" [ 882.977791] env[61006]: _type = "Task" [ 882.977791] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.985898] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.999200] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337146, 'name': PowerOffVM_Task, 'duration_secs': 0.256729} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.999451] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.999623] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.999910] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34d15f43-1d7d-4659-baf0-425a04f316d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.023868] env[61006]: DEBUG nova.compute.utils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.027027] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.027027] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.046851] env[61006]: DEBUG oslo_vmware.api [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22248} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.047120] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.047304] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.047479] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.047649] env[61006]: INFO nova.compute.manager [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Took 0.65 seconds to destroy the instance on the hypervisor. [ 883.047892] env[61006]: DEBUG oslo.service.loopingcall [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.048340] env[61006]: DEBUG nova.compute.manager [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.048439] env[61006]: DEBUG nova.network.neutron [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.065929] env[61006]: DEBUG nova.policy [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42027a2dce454ca98e295dadca850eee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2b5a4bf20a40ee9b8688935de75dc1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.079147] env[61006]: DEBUG nova.compute.manager [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Received event network-vif-plugged-0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.079371] env[61006]: DEBUG oslo_concurrency.lockutils [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] Acquiring lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.079647] env[61006]: DEBUG oslo_concurrency.lockutils [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.079881] env[61006]: DEBUG oslo_concurrency.lockutils [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.079990] env[61006]: DEBUG nova.compute.manager [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] No waiting events found dispatching network-vif-plugged-0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.080176] env[61006]: WARNING nova.compute.manager [req-434b76e1-22ba-4606-b70e-ba8af4cbc684 req-9cd75369-340c-493c-a606-15e2252fba8f service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Received unexpected event network-vif-plugged-0c39072b-7039-4a65-98e8-728f980a06b6 for instance with vm_state building and task_state spawning. [ 883.180903] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.181168] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.181353] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleting the datastore file [datastore2] cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.181637] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1895ee0-6423-4d3c-991b-fc0bb6b2e44f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.189959] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 883.189959] env[61006]: value = "task-1337150" [ 883.189959] env[61006]: _type = "Task" [ 883.189959] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.199930] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.248035] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525ec4e8-0fba-b90f-9c41-3ab15f802a1f, 'name': SearchDatastore_Task, 'duration_secs': 0.011007} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.252672] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.252993] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.253836] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.253991] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.254223] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.254505] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51bb1220-d6b0-4c84-9f92-cc8fd558d210 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.270020] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337144, 'name': CreateSnapshot_Task, 'duration_secs': 1.625804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.270020] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 883.270020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542ec521-e611-4667-bc74-6dd746548aec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.283186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.292046] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.292418] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.293626] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-192ec96d-d9e0-403b-93a3-a037116dc685 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.299944] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 883.299944] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d9d248-ba9f-db17-3fa0-32340b55701a" [ 883.299944] env[61006]: _type = "Task" [ 883.299944] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.310950] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d9d248-ba9f-db17-3fa0-32340b55701a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.476109] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Successfully created port: a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.493115] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037977} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.493115] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.493115] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.493115] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk to [datastore2] 2232c77d-9f4e-4981-9f55-614bba5d71db/2232c77d-9f4e-4981-9f55-614bba5d71db.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.493115] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fdc9b7a-cc4b-48b5-b43b-16da8258343b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.503171] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 883.503171] env[61006]: value = "task-1337151" [ 883.503171] env[61006]: _type = "Task" [ 883.503171] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.515113] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.519919] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Successfully updated port: 0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.528173] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating resource usage from migration 6d25c2c7-a5cf-4f93-b8eb-0de8b6265269 [ 883.531872] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.560451] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance cca6374c-09a4-4145-a116-c49c5a8330c1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.560645] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance e2a40bd9-fb66-40a2-bcf1-5c74707d59dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.560913] env[61006]: WARNING nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42b92d52-d1f0-48ff-94b6-6164b387456c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.561102] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 524f3fd1-1e71-40c0-96c2-0acac5055e01 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.561156] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0111452e-1b4f-499c-932d-f31364d1a14c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.561271] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance aedaa5d6-e0f2-492c-a14b-3254863e1f06 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.561415] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.561561] env[61006]: WARNING nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 9c9fa347-bcfe-4009-af72-5f427e9d234a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 883.561774] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d174a8ec-867e-4fea-b878-2a9af1476949 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.561984] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 66c93148-b91a-4d22-84af-f410c8e10875 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.562094] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.562225] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 2232c77d-9f4e-4981-9f55-614bba5d71db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.562363] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d1e09ace-177d-430a-9e65-d01368815272 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.562484] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 84befce1-ccd4-4937-9656-e5cbb7882f47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.562594] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance bdcf0445-6f8b-4e7b-bfb3-93a785162e56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 883.701408] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.800025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2290726c-520e-4fd6-962c-87a938d936f2 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-cca6374c-09a4-4145-a116-c49c5a8330c1-0df9528c-287c-45f7-8d8d-b43dc00c0b24" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.932s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.808196] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 883.808196] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2edf1874-2f2c-44e3-8968-52c48473ad58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.822773] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d9d248-ba9f-db17-3fa0-32340b55701a, 'name': SearchDatastore_Task, 'duration_secs': 0.086251} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.825564] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 883.825564] env[61006]: value = "task-1337152" [ 883.825564] env[61006]: _type = "Task" [ 883.825564] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.825803] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9984009-c891-4d9d-855a-6887649c5ddf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.836930] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 883.836930] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ea77f2-0bb8-77e8-1be4-e98c2b65a274" [ 883.836930] env[61006]: _type = "Task" [ 883.836930] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.839863] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.855146] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ea77f2-0bb8-77e8-1be4-e98c2b65a274, 'name': SearchDatastore_Task, 'duration_secs': 0.009806} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.855590] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.859031] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d1e09ace-177d-430a-9e65-d01368815272/d1e09ace-177d-430a-9e65-d01368815272.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.859031] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5b42133-791e-45bb-b211-9d084d8dd83d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.867257] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 883.867257] env[61006]: value = "task-1337153" [ 883.867257] env[61006]: _type = "Task" [ 883.867257] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.878749] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.012968] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.023035] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.023235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.023485] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.066687] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance c5d6f6df-d296-4d4d-815c-159dad2d08dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 884.066687] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Migration 6d25c2c7-a5cf-4f93-b8eb-0de8b6265269 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 884.066687] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 370f2153-adca-4513-8549-2bb7499cf913 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 884.166992] env[61006]: DEBUG nova.network.neutron [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.203405] env[61006]: DEBUG oslo_vmware.api [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.645861} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.203703] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.203900] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.204085] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.204267] env[61006]: INFO nova.compute.manager [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Took 1.74 seconds to destroy the instance on the hypervisor. [ 884.204512] env[61006]: DEBUG oslo.service.loopingcall [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.204722] env[61006]: DEBUG nova.compute.manager [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.204822] env[61006]: DEBUG nova.network.neutron [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.288399] env[61006]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 385eff2a-3efb-49b9-8063-c01a39357798 could not be found.", "detail": ""}} {{(pid=61006) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 884.288750] env[61006]: DEBUG nova.network.neutron [-] Unable to show port 385eff2a-3efb-49b9-8063-c01a39357798 as it no longer exists. {{(pid=61006) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 884.340103] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.378843] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.528656] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.541976] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.569174] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1a546f17-2fb8-4b99-9001-98cc6fe76837 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 884.578740] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.579063] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.579252] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.579502] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.579737] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.580392] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.580709] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.580904] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.581118] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.581405] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.581698] env[61006]: DEBUG nova.virt.hardware [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.582671] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc219797-62d5-4750-9d02-ad7197bcd0fc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.594373] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611f1aa7-1e15-49fa-918c-543e0b05f709 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.612829] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.670436] env[61006]: INFO nova.compute.manager [-] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Took 1.62 seconds to deallocate network for instance. [ 884.840750] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.881109] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.959221] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Updating instance_info_cache with network_info: [{"id": "0c39072b-7039-4a65-98e8-728f980a06b6", "address": "fa:16:3e:70:96:1d", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c39072b-70", "ovs_interfaceid": "0c39072b-7039-4a65-98e8-728f980a06b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.022489] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.072952] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance b27fc628-1ac8-4283-bf6e-bcff1cbfe149 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 885.073312] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 885.073469] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 885.117674] env[61006]: DEBUG nova.compute.manager [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Received event network-vif-plugged-a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.117907] env[61006]: DEBUG oslo_concurrency.lockutils [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] Acquiring lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.118966] env[61006]: DEBUG oslo_concurrency.lockutils [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.118966] env[61006]: DEBUG oslo_concurrency.lockutils [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.118966] env[61006]: DEBUG nova.compute.manager [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] No waiting events found dispatching network-vif-plugged-a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.118966] env[61006]: WARNING nova.compute.manager [req-99d233bc-b5f8-4091-a001-3181f0a641eb req-3488a804-9462-4408-acd6-6ec4cd50059a service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Received unexpected event network-vif-plugged-a7170e1b-5bc7-4be2-a529-25722de06d50 for instance with vm_state building and task_state spawning. [ 885.145504] env[61006]: DEBUG nova.compute.manager [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Received event network-changed-0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.145504] env[61006]: DEBUG nova.compute.manager [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Refreshing instance network info cache due to event network-changed-0c39072b-7039-4a65-98e8-728f980a06b6. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.145504] env[61006]: DEBUG oslo_concurrency.lockutils [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] Acquiring lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.178397] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.238457] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Successfully updated port: a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.351283] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.381821] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.415665] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e048c392-ef39-4ca8-ab2b-aef7d8bd1a2e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.426836] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d819681-e75f-4186-aad7-78da97d5c871 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.472978] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.473327] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Instance network_info: |[{"id": "0c39072b-7039-4a65-98e8-728f980a06b6", "address": "fa:16:3e:70:96:1d", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c39072b-70", "ovs_interfaceid": "0c39072b-7039-4a65-98e8-728f980a06b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.473942] env[61006]: DEBUG oslo_concurrency.lockutils [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] Acquired lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.474154] env[61006]: DEBUG nova.network.neutron [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Refreshing network info cache for port 0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.475466] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:96:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c39072b-7039-4a65-98e8-728f980a06b6', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.483046] env[61006]: DEBUG oslo.service.loopingcall [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.483767] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c476beda-abc3-4caf-a2ea-aa4cedb87f6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.487246] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.487495] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf5f76b6-ddce-40d8-b6de-85d1e895cf12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.509523] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffdb06d-93b9-4b0f-956a-2c9d344580a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.513785] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.513785] env[61006]: value = "task-1337154" [ 885.513785] env[61006]: _type = "Task" [ 885.513785] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.531849] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.537479] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.545750] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337154, 'name': CreateVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.672461] env[61006]: DEBUG nova.network.neutron [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.741316] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.741449] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.741624] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.844138] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.880961] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.029928] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337154, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.033227] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.040655] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.173987] env[61006]: INFO nova.compute.manager [-] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Took 1.97 seconds to deallocate network for instance. [ 886.224021] env[61006]: DEBUG nova.network.neutron [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Updated VIF entry in instance network info cache for port 0c39072b-7039-4a65-98e8-728f980a06b6. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.224480] env[61006]: DEBUG nova.network.neutron [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Updating instance_info_cache with network_info: [{"id": "0c39072b-7039-4a65-98e8-728f980a06b6", "address": "fa:16:3e:70:96:1d", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c39072b-70", "ovs_interfaceid": "0c39072b-7039-4a65-98e8-728f980a06b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.278064] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.344608] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.383378] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337153, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.494454} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.383378] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d1e09ace-177d-430a-9e65-d01368815272/d1e09ace-177d-430a-9e65-d01368815272.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.383378] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.383378] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1dd461a7-d3eb-4db7-a880-edf13d691380 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.390023] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 886.390023] env[61006]: value = "task-1337155" [ 886.390023] env[61006]: _type = "Task" [ 886.390023] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.397811] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.528982] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337151, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.797983} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.529364] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/06d8ddf9-1620-4813-97cd-7b51789db884/06d8ddf9-1620-4813-97cd-7b51789db884.vmdk to [datastore2] 2232c77d-9f4e-4981-9f55-614bba5d71db/2232c77d-9f4e-4981-9f55-614bba5d71db.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.530280] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c78f55-064c-4fb7-943c-6063bf6558dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.536209] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337154, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.548214] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 886.548481] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.038s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.557049] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 2232c77d-9f4e-4981-9f55-614bba5d71db/2232c77d-9f4e-4981-9f55-614bba5d71db.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.559709] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.474s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.559989] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.561818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.648s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.563283] env[61006]: INFO nova.compute.claims [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.565723] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33909503-e5e0-4a11-ab25-4cb1b896f9eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.586941] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 886.586941] env[61006]: value = "task-1337156" [ 886.586941] env[61006]: _type = "Task" [ 886.586941] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.596245] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.600375] env[61006]: INFO nova.scheduler.client.report [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Deleted allocations for instance 9c9fa347-bcfe-4009-af72-5f427e9d234a [ 886.680053] env[61006]: DEBUG nova.network.neutron [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Updating instance_info_cache with network_info: [{"id": "a7170e1b-5bc7-4be2-a529-25722de06d50", "address": "fa:16:3e:db:0b:29", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7170e1b-5b", "ovs_interfaceid": "a7170e1b-5bc7-4be2-a529-25722de06d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.681589] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.728189] env[61006]: DEBUG oslo_concurrency.lockutils [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] Releasing lock "refresh_cache-84befce1-ccd4-4937-9656-e5cbb7882f47" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.728523] env[61006]: DEBUG nova.compute.manager [req-cb992a22-465c-4749-a558-0a0657cd85b0 req-41b062cc-3219-41c4-a957-ce139b841a44 service nova] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Received event network-vif-deleted-5712d7c4-e719-4ef4-9d1c-b827b35dff58 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.841688] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.902032] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087518} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.902363] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.903247] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7082876c-8677-49cb-8819-b0435ea30807 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.928598] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] d1e09ace-177d-430a-9e65-d01368815272/d1e09ace-177d-430a-9e65-d01368815272.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.928932] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d822afc6-c15d-4506-a42a-9d3a51ea854b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.949829] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 886.949829] env[61006]: value = "task-1337157" [ 886.949829] env[61006]: _type = "Task" [ 886.949829] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.961154] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337157, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.031019] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337154, 'name': CreateVM_Task, 'duration_secs': 1.031639} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.031019] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.031019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.031019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.031019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.031019] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2250f4f9-407d-49aa-90f5-3686ac6f41fc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.036360] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 887.036360] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5274b051-08d1-8a40-7cf2-32dc3a882b57" [ 887.036360] env[61006]: _type = "Task" [ 887.036360] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.045246] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5274b051-08d1-8a40-7cf2-32dc3a882b57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.100019] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337156, 'name': ReconfigVM_Task, 'duration_secs': 0.342162} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.100019] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 2232c77d-9f4e-4981-9f55-614bba5d71db/2232c77d-9f4e-4981-9f55-614bba5d71db.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.100019] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7edcb42d-7f98-45ce-98b8-481809c18fc4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.109025] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 887.109025] env[61006]: value = "task-1337158" [ 887.109025] env[61006]: _type = "Task" [ 887.109025] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.112819] env[61006]: DEBUG oslo_concurrency.lockutils [None req-52bcb553-1f0c-40e7-bed8-5f49050f93a1 tempest-ServersNegativeTestMultiTenantJSON-55586113 tempest-ServersNegativeTestMultiTenantJSON-55586113-project-member] Lock "9c9fa347-bcfe-4009-af72-5f427e9d234a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.218s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.122829] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337158, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.186199] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.186199] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Instance network_info: |[{"id": "a7170e1b-5bc7-4be2-a529-25722de06d50", "address": "fa:16:3e:db:0b:29", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7170e1b-5b", "ovs_interfaceid": "a7170e1b-5bc7-4be2-a529-25722de06d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.186199] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:0b:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d413776-9a8c-4afd-856f-10dbb062ca95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7170e1b-5bc7-4be2-a529-25722de06d50', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.192297] env[61006]: DEBUG oslo.service.loopingcall [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.192760] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.193107] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-023f8d14-f278-4515-9127-c39ddb738f31 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.214634] env[61006]: DEBUG nova.compute.manager [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Received event network-changed-a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.215116] env[61006]: DEBUG nova.compute.manager [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Refreshing instance network info cache due to event network-changed-a7170e1b-5bc7-4be2-a529-25722de06d50. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.215460] env[61006]: DEBUG oslo_concurrency.lockutils [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] Acquiring lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.215724] env[61006]: DEBUG oslo_concurrency.lockutils [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] Acquired lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.215998] env[61006]: DEBUG nova.network.neutron [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Refreshing network info cache for port a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.217943] env[61006]: DEBUG nova.compute.manager [req-b72a6cec-0c84-4172-9f12-696b51d8d221 req-a22f1e78-bda2-40ce-b918-aa7ca4b5703b service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-deleted-598c1248-f434-4f05-ac91-c7f1322879fa {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.232019] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.232019] env[61006]: value = "task-1337159" [ 887.232019] env[61006]: _type = "Task" [ 887.232019] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.240104] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337159, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.345838] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337152, 'name': CloneVM_Task, 'duration_secs': 3.084175} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.346232] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created linked-clone VM from snapshot [ 887.347116] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32394650-fb72-4c6e-bd77-0125d1325ec2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.356833] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploading image e72dabda-ebc9-4e91-a9b2-f12e7a212e79 {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 887.380336] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.380336] env[61006]: value = "vm-285362" [ 887.380336] env[61006]: _type = "VirtualMachine" [ 887.380336] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.380759] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6a575255-499d-4f81-915e-c5a68bb53353 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.388818] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease: (returnval){ [ 887.388818] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5254e0ef-9937-0357-3774-5cd554988955" [ 887.388818] env[61006]: _type = "HttpNfcLease" [ 887.388818] env[61006]: } obtained for exporting VM: (result){ [ 887.388818] env[61006]: value = "vm-285362" [ 887.388818] env[61006]: _type = "VirtualMachine" [ 887.388818] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.389252] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the lease: (returnval){ [ 887.389252] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5254e0ef-9937-0357-3774-5cd554988955" [ 887.389252] env[61006]: _type = "HttpNfcLease" [ 887.389252] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.396720] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.396720] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5254e0ef-9937-0357-3774-5cd554988955" [ 887.396720] env[61006]: _type = "HttpNfcLease" [ 887.396720] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.460334] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337157, 'name': ReconfigVM_Task, 'duration_secs': 0.306534} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.460671] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Reconfigured VM instance instance-00000051 to attach disk [datastore2] d1e09ace-177d-430a-9e65-d01368815272/d1e09ace-177d-430a-9e65-d01368815272.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.461378] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6fbd4c4-b43d-4096-84b3-13fb71722f3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.468597] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 887.468597] env[61006]: value = "task-1337161" [ 887.468597] env[61006]: _type = "Task" [ 887.468597] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.477286] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337161, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.549422] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5274b051-08d1-8a40-7cf2-32dc3a882b57, 'name': SearchDatastore_Task, 'duration_secs': 0.010312} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.549815] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.550161] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.550454] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.550612] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.550858] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.551166] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbb48435-da6d-44c3-8759-7ae6c14aba02 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.561742] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.561964] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.562871] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70e8ddfd-270e-4646-90b5-d51aab85ad8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.569521] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 887.569521] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52da94af-a329-79e4-b81e-13b37c5a516a" [ 887.569521] env[61006]: _type = "Task" [ 887.569521] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.579718] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52da94af-a329-79e4-b81e-13b37c5a516a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.618280] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337158, 'name': Rename_Task, 'duration_secs': 0.173685} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.618702] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.618802] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a89ab540-cbab-41e7-b4e1-00361361dbbc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.626265] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 887.626265] env[61006]: value = "task-1337162" [ 887.626265] env[61006]: _type = "Task" [ 887.626265] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.637750] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.743338] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337159, 'name': CreateVM_Task, 'duration_secs': 0.454385} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.743338] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.743338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.743338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.743338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.743782] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c3f5c5d-fc87-4dfa-992f-c6082e058937 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.749654] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 887.749654] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294da85-b3a2-c400-df87-09bffdd6c29e" [ 887.749654] env[61006]: _type = "Task" [ 887.749654] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.763915] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294da85-b3a2-c400-df87-09bffdd6c29e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.881635] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d277959-760a-4b52-9063-5cce28779d58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.890078] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75db0159-10ca-47a0-8255-bb5bb9ac43cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.902531] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.902531] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5254e0ef-9937-0357-3774-5cd554988955" [ 887.902531] env[61006]: _type = "HttpNfcLease" [ 887.902531] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 887.930996] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 887.930996] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5254e0ef-9937-0357-3774-5cd554988955" [ 887.930996] env[61006]: _type = "HttpNfcLease" [ 887.930996] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 887.932427] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcb9530-5f9f-49bf-9e94-fe66a07383a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.938222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632bb71e-8bd4-4a75-afdd-8cc1d0a3936a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.951064] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 887.951270] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 887.953637] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36f8062-c433-4bc2-ab5d-ccaec8e72696 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.020937] env[61006]: DEBUG nova.network.neutron [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Updated VIF entry in instance network info cache for port a7170e1b-5bc7-4be2-a529-25722de06d50. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.021309] env[61006]: DEBUG nova.network.neutron [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Updating instance_info_cache with network_info: [{"id": "a7170e1b-5bc7-4be2-a529-25722de06d50", "address": "fa:16:3e:db:0b:29", "network": {"id": "6dc9026d-d661-4252-ac10-d53c6d74bd67", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1165197363-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2b5a4bf20a40ee9b8688935de75dc1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d413776-9a8c-4afd-856f-10dbb062ca95", "external-id": "nsx-vlan-transportzone-913", "segmentation_id": 913, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7170e1b-5b", "ovs_interfaceid": "a7170e1b-5bc7-4be2-a529-25722de06d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.026163] env[61006]: DEBUG nova.compute.provider_tree [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.036093] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337161, 'name': Rename_Task, 'duration_secs': 0.19465} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.036946] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.037732] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e3241d3-3b8f-4c68-9eb7-ce7aced43760 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.045672] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 888.045672] env[61006]: value = "task-1337163" [ 888.045672] env[61006]: _type = "Task" [ 888.045672] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.055083] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.068231] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c0c1c4cf-7743-4407-b840-976c0270d577 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.082408] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52da94af-a329-79e4-b81e-13b37c5a516a, 'name': SearchDatastore_Task, 'duration_secs': 0.012334} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.082824] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-783a47fe-5fb9-42bb-b78a-00b4a258e68c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.090126] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 888.090126] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520343a8-217c-9e0f-07c7-e0797d2242a8" [ 888.090126] env[61006]: _type = "Task" [ 888.090126] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.102810] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520343a8-217c-9e0f-07c7-e0797d2242a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.138030] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337162, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.264744] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294da85-b3a2-c400-df87-09bffdd6c29e, 'name': SearchDatastore_Task, 'duration_secs': 0.027433} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.265049] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.265381] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.265690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.528537] env[61006]: DEBUG oslo_concurrency.lockutils [req-e85b8d0a-353a-4509-9d54-50baa3a5220e req-731dfdac-6bf6-4418-8d71-a84d3b22e60e service nova] Releasing lock "refresh_cache-bdcf0445-6f8b-4e7b-bfb3-93a785162e56" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.537997] env[61006]: DEBUG nova.scheduler.client.report [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.560537] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337163, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.606676] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520343a8-217c-9e0f-07c7-e0797d2242a8, 'name': SearchDatastore_Task, 'duration_secs': 0.018205} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.606730] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.607157] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 84befce1-ccd4-4937-9656-e5cbb7882f47/84befce1-ccd4-4937-9656-e5cbb7882f47.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.607639] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.607968] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.608372] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc902de2-7a9d-4d97-b7c1-c467d6ed5af0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.610757] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10997a23-a7b6-4029-8d9b-478f88c110b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.618980] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 888.618980] env[61006]: value = "task-1337164" [ 888.618980] env[61006]: _type = "Task" [ 888.618980] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.624339] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.624714] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.626023] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23a0b758-f05b-4a69-9fdd-e286c5bda195 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.632740] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337164, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.641379] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 888.641379] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524ce623-1795-ec87-711f-4a6cb128bcdc" [ 888.641379] env[61006]: _type = "Task" [ 888.641379] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.645593] env[61006]: DEBUG oslo_vmware.api [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337162, 'name': PowerOnVM_Task, 'duration_secs': 0.872369} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.649682] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.650120] env[61006]: INFO nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Took 17.72 seconds to spawn the instance on the hypervisor. [ 888.650436] env[61006]: DEBUG nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.651441] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c46eeb-e60e-43d0-b3a6-bef668c93762 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.666699] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524ce623-1795-ec87-711f-4a6cb128bcdc, 'name': SearchDatastore_Task, 'duration_secs': 0.012804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.667824] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f61202d8-7ee0-467d-822e-5d7ae78dea53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.675317] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 888.675317] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fddb55-e5d3-cdd0-67f6-ffe84c2a747b" [ 888.675317] env[61006]: _type = "Task" [ 888.675317] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.686646] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fddb55-e5d3-cdd0-67f6-ffe84c2a747b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.045892] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.047014] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.053023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 17.474s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.063079] env[61006]: DEBUG oslo_vmware.api [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337163, 'name': PowerOnVM_Task, 'duration_secs': 0.599275} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.063386] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.063667] env[61006]: INFO nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Took 11.71 seconds to spawn the instance on the hypervisor. [ 889.066117] env[61006]: DEBUG nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.066117] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9dd87d-4436-4425-bfb4-030a69b02b07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.134031] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337164, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.174463] env[61006]: INFO nova.compute.manager [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Took 43.86 seconds to build instance. [ 889.186889] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52fddb55-e5d3-cdd0-67f6-ffe84c2a747b, 'name': SearchDatastore_Task, 'duration_secs': 0.013257} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.187204] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.187465] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bdcf0445-6f8b-4e7b-bfb3-93a785162e56/bdcf0445-6f8b-4e7b-bfb3-93a785162e56.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.187790] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e1d476d-df93-4936-bae9-b84283f35c26 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.198303] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 889.198303] env[61006]: value = "task-1337165" [ 889.198303] env[61006]: _type = "Task" [ 889.198303] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.208228] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.556410] env[61006]: INFO nova.compute.claims [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.568015] env[61006]: DEBUG nova.compute.utils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.573892] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 889.574101] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.595878] env[61006]: INFO nova.compute.manager [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Took 30.87 seconds to build instance. [ 889.632506] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337164, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641312} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.634293] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 84befce1-ccd4-4937-9656-e5cbb7882f47/84befce1-ccd4-4937-9656-e5cbb7882f47.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.634293] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.635395] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0af9216-adc6-4787-b8d9-b9b9507bc21a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.644496] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 889.644496] env[61006]: value = "task-1337166" [ 889.644496] env[61006]: _type = "Task" [ 889.644496] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.650926] env[61006]: DEBUG nova.policy [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f5eec5c1855d4ce091bbb7be3c811739', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3ab42ad314646e792f4ad39bfc40cf8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.661075] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.676983] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a515ca17-83b4-403f-aa62-7a2abaf1e57b tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.844s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.708958] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337165, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.720667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "2232c77d-9f4e-4981-9f55-614bba5d71db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.720667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.720667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.720667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.721555] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.726886] env[61006]: INFO nova.compute.manager [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Terminating instance [ 889.728079] env[61006]: DEBUG nova.compute.manager [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.729052] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.729694] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b4bcda-4025-4611-a6f9-113d604018d4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.738728] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.739267] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fae7a35-bd98-45d7-a060-8b6a69a71c90 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.749240] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 889.749240] env[61006]: value = "task-1337167" [ 889.749240] env[61006]: _type = "Task" [ 889.749240] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.758308] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.076817] env[61006]: INFO nova.compute.resource_tracker [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating resource usage from migration 6d25c2c7-a5cf-4f93-b8eb-0de8b6265269 [ 890.083739] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.100701] env[61006]: DEBUG oslo_concurrency.lockutils [None req-08420a74-8462-4097-8327-538bd6351897 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.385s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.139777] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Successfully created port: 2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.156016] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235687} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.160091] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.161849] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb391e0f-6a02-48bb-9bf0-9552add1a588 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.188618] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 84befce1-ccd4-4937-9656-e5cbb7882f47/84befce1-ccd4-4937-9656-e5cbb7882f47.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.192248] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c6562f6-2a67-4ff4-ae97-01f3bee4678e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.226021] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.735146} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.228187] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bdcf0445-6f8b-4e7b-bfb3-93a785162e56/bdcf0445-6f8b-4e7b-bfb3-93a785162e56.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.228635] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.228849] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 890.228849] env[61006]: value = "task-1337168" [ 890.228849] env[61006]: _type = "Task" [ 890.228849] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.229365] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1e5e733-72ac-4fac-9420-986fb9452cb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.241802] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337168, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.243587] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 890.243587] env[61006]: value = "task-1337169" [ 890.243587] env[61006]: _type = "Task" [ 890.243587] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.261285] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337167, 'name': PowerOffVM_Task, 'duration_secs': 0.459646} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.264161] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.264161] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.264161] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.266469] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-769ebb94-ee4e-4ec3-9482-a17cc772e684 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.334113] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.334342] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.334485] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleting the datastore file [datastore2] 2232c77d-9f4e-4981-9f55-614bba5d71db {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.334757] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ddc913b-d01b-47fa-b128-514cce3d70f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.342418] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 890.342418] env[61006]: value = "task-1337171" [ 890.342418] env[61006]: _type = "Task" [ 890.342418] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.354414] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.433362] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d1e09ace-177d-430a-9e65-d01368815272" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.433721] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.433956] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d1e09ace-177d-430a-9e65-d01368815272-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.434177] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.434508] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.436629] env[61006]: INFO nova.compute.manager [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Terminating instance [ 890.439384] env[61006]: DEBUG nova.compute.manager [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.439384] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.439734] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672fb5a7-5e60-4b9b-a0d0-28b34abffd86 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.454026] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.454166] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-119fc8e1-c707-4afa-b00a-124333c2b57b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.462794] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 890.462794] env[61006]: value = "task-1337172" [ 890.462794] env[61006]: _type = "Task" [ 890.462794] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.471559] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337172, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.479211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3310ac74-c56c-4198-b38a-7ee2799c9cc3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.487632] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b283efe-7864-41a8-820f-606fa06dfbd8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.521115] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de261a7-b531-4c74-9207-60791a508896 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.530354] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6933a514-9787-492f-8d37-973ca3e4fdab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.548795] env[61006]: DEBUG nova.compute.provider_tree [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.742293] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337168, 'name': ReconfigVM_Task, 'duration_secs': 0.50947} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.745167] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 84befce1-ccd4-4937-9656-e5cbb7882f47/84befce1-ccd4-4937-9656-e5cbb7882f47.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.745167] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be264415-4c17-482a-82f5-f020d4bdc675 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.757300] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126139} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.759082] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.759674] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 890.759674] env[61006]: value = "task-1337173" [ 890.759674] env[61006]: _type = "Task" [ 890.759674] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.761262] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1eea2c-079e-44f9-8999-5aab7e4753b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.791097] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] bdcf0445-6f8b-4e7b-bfb3-93a785162e56/bdcf0445-6f8b-4e7b-bfb3-93a785162e56.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.795070] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd5d9aad-5fe0-4fad-841c-ca6dd56fb7d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.809405] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337173, 'name': Rename_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.814185] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 890.814185] env[61006]: value = "task-1337174" [ 890.814185] env[61006]: _type = "Task" [ 890.814185] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.823263] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337174, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.856237] env[61006]: DEBUG oslo_vmware.api [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295392} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.856561] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.856895] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.857153] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.857425] env[61006]: INFO nova.compute.manager [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Took 1.13 seconds to destroy the instance on the hypervisor. [ 890.857732] env[61006]: DEBUG oslo.service.loopingcall [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.858023] env[61006]: DEBUG nova.compute.manager [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.858161] env[61006]: DEBUG nova.network.neutron [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.973923] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337172, 'name': PowerOffVM_Task, 'duration_secs': 0.228483} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.974215] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.974462] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.974787] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b812f475-711f-46c1-af16-a76222a8db4a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.050658] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.050658] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.050658] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore2] d1e09ace-177d-430a-9e65-d01368815272 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.050658] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01b31228-de86-4211-a9e7-d0f3ff80db43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.053401] env[61006]: DEBUG nova.scheduler.client.report [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.058423] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 891.058423] env[61006]: value = "task-1337176" [ 891.058423] env[61006]: _type = "Task" [ 891.058423] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.072267] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.099621] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.130363] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.130659] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.130822] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.131123] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.131395] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.131640] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.131971] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.132248] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.132523] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.132796] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.133098] env[61006]: DEBUG nova.virt.hardware [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.134281] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6f9aa3-d61f-4d82-a7da-5f967dd183ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.145288] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6308f22-7290-42f5-9f54-d33575386e17 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.195801] env[61006]: DEBUG nova.compute.manager [req-fdf14f79-319a-4a1e-b251-e6db1dab7676 req-d8371e57-0a6b-45bf-bcb2-d15836c3f5ea service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Received event network-vif-deleted-8cc8c41b-62f9-4046-a1c9-764067e78608 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.196486] env[61006]: INFO nova.compute.manager [req-fdf14f79-319a-4a1e-b251-e6db1dab7676 req-d8371e57-0a6b-45bf-bcb2-d15836c3f5ea service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Neutron deleted interface 8cc8c41b-62f9-4046-a1c9-764067e78608; detaching it from the instance and deleting it from the info cache [ 891.196795] env[61006]: DEBUG nova.network.neutron [req-fdf14f79-319a-4a1e-b251-e6db1dab7676 req-d8371e57-0a6b-45bf-bcb2-d15836c3f5ea service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.278811] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337173, 'name': Rename_Task, 'duration_secs': 0.151077} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.279061] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.279489] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b2e0be7-825c-43b8-8503-d109a7b08b51 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.286285] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 891.286285] env[61006]: value = "task-1337177" [ 891.286285] env[61006]: _type = "Task" [ 891.286285] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.294738] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.324798] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337174, 'name': ReconfigVM_Task, 'duration_secs': 0.481602} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.325185] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Reconfigured VM instance instance-00000053 to attach disk [datastore2] bdcf0445-6f8b-4e7b-bfb3-93a785162e56/bdcf0445-6f8b-4e7b-bfb3-93a785162e56.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.325866] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56bd1f11-fc00-45f1-a5b4-af236455adbc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.332130] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 891.332130] env[61006]: value = "task-1337178" [ 891.332130] env[61006]: _type = "Task" [ 891.332130] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.341029] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337178, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.560603] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.511s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.561236] env[61006]: INFO nova.compute.manager [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Migrating [ 891.561236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.561582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.562860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.601s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.563078] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.565246] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.130s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.566782] env[61006]: INFO nova.compute.claims [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.580129] env[61006]: DEBUG oslo_vmware.api [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.297579} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.580470] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.580605] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.580786] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.580961] env[61006]: INFO nova.compute.manager [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d1e09ace-177d-430a-9e65-d01368815272] Took 1.14 seconds to destroy the instance on the hypervisor. [ 891.581445] env[61006]: DEBUG oslo.service.loopingcall [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.581445] env[61006]: DEBUG nova.compute.manager [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.581532] env[61006]: DEBUG nova.network.neutron [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 891.591258] env[61006]: DEBUG nova.network.neutron [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.598728] env[61006]: INFO nova.scheduler.client.report [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Deleted allocations for instance 42b92d52-d1f0-48ff-94b6-6164b387456c [ 891.659435] env[61006]: DEBUG nova.compute.manager [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Received event network-vif-plugged-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.659760] env[61006]: DEBUG oslo_concurrency.lockutils [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] Acquiring lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.659927] env[61006]: DEBUG oslo_concurrency.lockutils [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.660235] env[61006]: DEBUG oslo_concurrency.lockutils [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.660341] env[61006]: DEBUG nova.compute.manager [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] No waiting events found dispatching network-vif-plugged-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 891.660502] env[61006]: WARNING nova.compute.manager [req-225f5ac2-d76a-4273-90f4-bbbd12688d64 req-b78c944f-a8e1-4d32-bec3-94bfb48b48c3 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Received unexpected event network-vif-plugged-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 for instance with vm_state building and task_state spawning. [ 891.700684] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47503458-1768-4f09-ae7f-4dbbad7deb0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.709730] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05e0806-0eb5-461f-9719-0e3c9a5ce337 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.744811] env[61006]: DEBUG nova.compute.manager [req-fdf14f79-319a-4a1e-b251-e6db1dab7676 req-d8371e57-0a6b-45bf-bcb2-d15836c3f5ea service nova] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Detach interface failed, port_id=8cc8c41b-62f9-4046-a1c9-764067e78608, reason: Instance 2232c77d-9f4e-4981-9f55-614bba5d71db could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 891.802032] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337177, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.841486] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337178, 'name': Rename_Task, 'duration_secs': 0.203408} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.841866] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.842267] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b770c5c-8139-42ed-ac89-14e894f804f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.849139] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 891.849139] env[61006]: value = "task-1337179" [ 891.849139] env[61006]: _type = "Task" [ 891.849139] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.863760] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.864735] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Successfully updated port: 2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.074484] env[61006]: INFO nova.compute.rpcapi [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 892.075140] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.095129] env[61006]: INFO nova.compute.manager [-] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Took 1.24 seconds to deallocate network for instance. [ 892.106051] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5a385ea1-408e-42c2-8b7f-b42639a52607 tempest-ServersTestManualDisk-1208872639 tempest-ServersTestManualDisk-1208872639-project-member] Lock "42b92d52-d1f0-48ff-94b6-6164b387456c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.922s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.300259] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337177, 'name': PowerOnVM_Task, 'duration_secs': 0.530869} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.303089] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.303351] env[61006]: INFO nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Took 10.36 seconds to spawn the instance on the hypervisor. [ 892.303661] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.304775] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9789892e-cd92-47ee-a2fd-3ca7e489a013 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.342248] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d34e8ad-c444-4e3b-a9db-a7f161f20dc2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.350863] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c625f6bc-fe08-4ad1-894c-3074d2af62ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.363611] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337179, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.391167] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.391354] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquired lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.392394] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.395037] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af28ec7d-17c9-48b0-a1d2-91f534598bf0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.402422] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63df868e-18b6-45c6-aa7f-05485a1a1403 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.423386] env[61006]: DEBUG nova.compute.provider_tree [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.444494] env[61006]: DEBUG nova.network.neutron [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.592293] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.592512] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.592712] env[61006]: DEBUG nova.network.neutron [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.601994] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.823849] env[61006]: INFO nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Took 30.31 seconds to build instance. [ 892.863216] env[61006]: DEBUG oslo_vmware.api [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337179, 'name': PowerOnVM_Task, 'duration_secs': 0.811612} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.863441] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.863644] env[61006]: INFO nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Took 8.32 seconds to spawn the instance on the hypervisor. [ 892.863824] env[61006]: DEBUG nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.864656] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eec3931-9f6b-43fc-804f-1ad258f54f30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.926403] env[61006]: DEBUG nova.scheduler.client.report [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.947275] env[61006]: INFO nova.compute.manager [-] [instance: d1e09ace-177d-430a-9e65-d01368815272] Took 1.37 seconds to deallocate network for instance. [ 892.967048] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.154728] env[61006]: DEBUG nova.network.neutron [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updating instance_info_cache with network_info: [{"id": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "address": "fa:16:3e:a9:6f:95", "network": {"id": "6bbe0858-571f-4b0c-8912-063ff6b5469c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-740305265-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3ab42ad314646e792f4ad39bfc40cf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f257dd0-56", "ovs_interfaceid": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.224558] env[61006]: DEBUG nova.compute.manager [req-98bd12e5-3761-4042-9bdb-524578dd967f req-fd4ea6da-1aa6-4c15-a44c-be519a2d4e3f service nova] [instance: d1e09ace-177d-430a-9e65-d01368815272] Received event network-vif-deleted-d6d7cece-1280-46c8-ad3b-68ec36e57ebf {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.325729] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.840s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.381709] env[61006]: INFO nova.compute.manager [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Took 30.84 seconds to build instance. [ 893.431298] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.866s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.431979] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.434689] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.690s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.436183] env[61006]: INFO nova.compute.claims [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.454551] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.505852] env[61006]: DEBUG nova.network.neutron [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.656251] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Releasing lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.656529] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Instance network_info: |[{"id": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "address": "fa:16:3e:a9:6f:95", "network": {"id": "6bbe0858-571f-4b0c-8912-063ff6b5469c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-740305265-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3ab42ad314646e792f4ad39bfc40cf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f257dd0-56", "ovs_interfaceid": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.657077] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:6f:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c66a277b-e3bf-43b8-a632-04fdd0720b91', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.665060] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Creating folder: Project (a3ab42ad314646e792f4ad39bfc40cf8). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.665379] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-061fb2fe-4542-4f5b-9a53-c49b4ecb39c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.675955] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Created folder: Project (a3ab42ad314646e792f4ad39bfc40cf8) in parent group-v285275. [ 893.676202] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Creating folder: Instances. Parent ref: group-v285365. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.676448] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47c87333-e873-4e79-97c0-9a4e0341f1cf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.686648] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Created folder: Instances in parent group-v285365. [ 893.686906] env[61006]: DEBUG oslo.service.loopingcall [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.687123] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.687344] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15aa1d5a-3c5a-47f2-8de0-f00c2f454e1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.704730] env[61006]: DEBUG nova.compute.manager [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Received event network-changed-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.704938] env[61006]: DEBUG nova.compute.manager [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Refreshing instance network info cache due to event network-changed-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 893.705170] env[61006]: DEBUG oslo_concurrency.lockutils [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] Acquiring lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.705321] env[61006]: DEBUG oslo_concurrency.lockutils [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] Acquired lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.705500] env[61006]: DEBUG nova.network.neutron [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Refreshing network info cache for port 2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.712176] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.712176] env[61006]: value = "task-1337182" [ 893.712176] env[61006]: _type = "Task" [ 893.712176] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.721638] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337182, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.884160] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0ee21512-513d-4517-87e7-461393e957ff tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.356s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.945200] env[61006]: DEBUG nova.compute.utils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.952778] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.953138] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 894.008734] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.027154] env[61006]: DEBUG nova.policy [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc2220b969c942efba38ebd09ec63ed6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbbb1cf0f7974ccdb451078aaa448272', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.223588] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337182, 'name': CreateVM_Task, 'duration_secs': 0.367423} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.223588] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.224103] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.224275] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.224829] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.225136] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcdb8785-3cb8-4ba9-aab3-54fa6349e703 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.231069] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 894.231069] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525f27af-f67b-42cf-75fa-f6084d5e4277" [ 894.231069] env[61006]: _type = "Task" [ 894.231069] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.240022] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525f27af-f67b-42cf-75fa-f6084d5e4277, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.460594] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.647380] env[61006]: DEBUG nova.network.neutron [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updated VIF entry in instance network info cache for port 2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.647755] env[61006]: DEBUG nova.network.neutron [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updating instance_info_cache with network_info: [{"id": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "address": "fa:16:3e:a9:6f:95", "network": {"id": "6bbe0858-571f-4b0c-8912-063ff6b5469c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-740305265-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3ab42ad314646e792f4ad39bfc40cf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f257dd0-56", "ovs_interfaceid": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.733036] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Successfully created port: 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.747575] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525f27af-f67b-42cf-75fa-f6084d5e4277, 'name': SearchDatastore_Task, 'duration_secs': 0.017821} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.747858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.748131] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.749784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.749784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.749784] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.750082] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3eb53f4e-324b-4202-befa-18218a0a0fc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.764490] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.764688] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.765505] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e6f503c-eef5-4832-ac12-f7a3cbf5f757 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.771153] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 894.771153] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522e133b-abea-3a19-03b6-30f8a08ba4dc" [ 894.771153] env[61006]: _type = "Task" [ 894.771153] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.784325] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522e133b-abea-3a19-03b6-30f8a08ba4dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.825389] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4360397f-bd4e-4bd9-ad67-b68423acd392 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.841925] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dae016b-b6e3-4d14-982f-ae16fafcb9a1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.891401] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dc00be-2f2b-4838-97dd-a7b3d4349c4f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.902490] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6210b5ef-6d67-419f-ada7-287c3dee317e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.927444] env[61006]: DEBUG nova.compute.provider_tree [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.063531] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "84befce1-ccd4-4937-9656-e5cbb7882f47" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.064409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.064409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.064409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.064619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.067392] env[61006]: INFO nova.compute.manager [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Terminating instance [ 895.070708] env[61006]: DEBUG nova.compute.manager [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 895.070708] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.070708] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f88c03a-8751-406f-a36c-d207e54b5d81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.078211] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.078482] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7a564c1-ea28-417e-93f8-4e9e0f3fb9fc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.085500] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 895.085500] env[61006]: value = "task-1337183" [ 895.085500] env[61006]: _type = "Task" [ 895.085500] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.095942] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.153421] env[61006]: DEBUG oslo_concurrency.lockutils [req-009c692b-6398-4517-a30e-97624f3c8d46 req-e66c4aa9-29bf-442c-8e64-94e5b1677951 service nova] Releasing lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.286274] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522e133b-abea-3a19-03b6-30f8a08ba4dc, 'name': SearchDatastore_Task, 'duration_secs': 0.01417} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.287237] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c34431c6-c12d-4fe7-a7f2-a2b9280f461a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.293622] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 895.293622] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf67fc-f574-e3fb-ba7c-37aebe21e470" [ 895.293622] env[61006]: _type = "Task" [ 895.293622] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.302385] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf67fc-f574-e3fb-ba7c-37aebe21e470, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.319112] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.319533] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.319867] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.320203] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.320422] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.323057] env[61006]: INFO nova.compute.manager [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Terminating instance [ 895.326989] env[61006]: DEBUG nova.compute.manager [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 895.327261] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.328401] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141528ed-6c40-4faf-8f52-eb95e417cd30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.336218] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.336476] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79339216-bb2f-4931-bbe7-b93e25766f3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.343741] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 895.343741] env[61006]: value = "task-1337184" [ 895.343741] env[61006]: _type = "Task" [ 895.343741] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.352026] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.431335] env[61006]: DEBUG nova.scheduler.client.report [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.478079] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.502782] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.503093] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.503214] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.503399] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.503662] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.503882] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.505044] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.505044] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.505044] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.505044] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.505044] env[61006]: DEBUG nova.virt.hardware [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.505701] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962ac422-7c90-493b-8f8a-c81d80cf90a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.513963] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d2edaa-78b0-4531-9ce0-262f26716363 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.534129] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af827adc-5c23-4ece-aaea-6d2607343e1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.550793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 0 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 895.596055] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337183, 'name': PowerOffVM_Task, 'duration_secs': 0.257613} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.596353] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.596775] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.597979] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13577574-c06d-47b6-9a15-ed4bec620c36 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.658540] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.658793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.658973] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleting the datastore file [datastore2] 84befce1-ccd4-4937-9656-e5cbb7882f47 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.659267] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64c8d25a-6491-490c-8c2f-228afcc1a09c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.667033] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 895.667033] env[61006]: value = "task-1337186" [ 895.667033] env[61006]: _type = "Task" [ 895.667033] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.675658] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.805232] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf67fc-f574-e3fb-ba7c-37aebe21e470, 'name': SearchDatastore_Task, 'duration_secs': 0.012873} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.805526] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.805792] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] c5d6f6df-d296-4d4d-815c-159dad2d08dc/c5d6f6df-d296-4d4d-815c-159dad2d08dc.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 895.806080] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e44ca203-7458-4543-b153-c712ae280c27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.812761] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 895.812761] env[61006]: value = "task-1337187" [ 895.812761] env[61006]: _type = "Task" [ 895.812761] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.822330] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.853444] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337184, 'name': PowerOffVM_Task, 'duration_secs': 0.193228} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.853943] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.853943] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.854158] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-842219e6-a65d-4478-a758-bf7576adb0ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.928531] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.928886] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.929039] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleting the datastore file [datastore2] bdcf0445-6f8b-4e7b-bfb3-93a785162e56 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.929268] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93781b93-048a-4110-99cb-783270f6f4ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.936141] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.936648] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.940185] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for the task: (returnval){ [ 895.940185] env[61006]: value = "task-1337189" [ 895.940185] env[61006]: _type = "Task" [ 895.940185] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.940185] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.762s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.940185] env[61006]: DEBUG nova.objects.instance [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lazy-loading 'resources' on Instance uuid 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.949538] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.057128] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.057484] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af163403-9450-4950-bfca-a532ac6d1765 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.065491] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 896.065491] env[61006]: value = "task-1337190" [ 896.065491] env[61006]: _type = "Task" [ 896.065491] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.076314] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.180995] env[61006]: DEBUG oslo_vmware.api [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165255} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.181347] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.181540] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.183552] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.183552] env[61006]: INFO nova.compute.manager [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Took 1.11 seconds to destroy the instance on the hypervisor. [ 896.183552] env[61006]: DEBUG oslo.service.loopingcall [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.183552] env[61006]: DEBUG nova.compute.manager [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.183552] env[61006]: DEBUG nova.network.neutron [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.323638] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337187, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.441648] env[61006]: DEBUG nova.compute.utils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.443289] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.443366] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.458023] env[61006]: DEBUG oslo_vmware.api [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Task: {'id': task-1337189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.485075} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.458772] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.459561] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.459561] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.459561] env[61006]: INFO nova.compute.manager [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Took 1.13 seconds to destroy the instance on the hypervisor. [ 896.459722] env[61006]: DEBUG oslo.service.loopingcall [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.460899] env[61006]: DEBUG nova.compute.manager [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.460899] env[61006]: DEBUG nova.network.neutron [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.466320] env[61006]: DEBUG nova.compute.manager [req-382ec1e6-ad7a-469f-8be1-4d032afaa2ba req-00bdd7e6-b017-4915-ac55-e3d86cbc8474 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Received event network-vif-deleted-0c39072b-7039-4a65-98e8-728f980a06b6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.466320] env[61006]: INFO nova.compute.manager [req-382ec1e6-ad7a-469f-8be1-4d032afaa2ba req-00bdd7e6-b017-4915-ac55-e3d86cbc8474 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Neutron deleted interface 0c39072b-7039-4a65-98e8-728f980a06b6; detaching it from the instance and deleting it from the info cache [ 896.467446] env[61006]: DEBUG nova.network.neutron [req-382ec1e6-ad7a-469f-8be1-4d032afaa2ba req-00bdd7e6-b017-4915-ac55-e3d86cbc8474 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.483644] env[61006]: DEBUG nova.compute.manager [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-vif-plugged-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.483644] env[61006]: DEBUG oslo_concurrency.lockutils [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.483644] env[61006]: DEBUG oslo_concurrency.lockutils [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.484033] env[61006]: DEBUG oslo_concurrency.lockutils [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.484033] env[61006]: DEBUG nova.compute.manager [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] No waiting events found dispatching network-vif-plugged-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.484145] env[61006]: WARNING nova.compute.manager [req-6931ac39-502f-48d2-8792-8bda0226b2ef req-98c32069-28fd-49d3-8034-74556c667ca8 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received unexpected event network-vif-plugged-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f for instance with vm_state building and task_state spawning. [ 896.515667] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 896.517170] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed91851b-1bb6-4665-b776-ad03321e6f27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.527286] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 896.527447] env[61006]: ERROR oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk due to incomplete transfer. [ 896.527678] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c163098b-9a6b-456a-a854-23dee6128b66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.532329] env[61006]: DEBUG nova.policy [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '984ecb8fe25e439a9eceb7c3486c4663', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e28ff2841a84b969a83d6b63f621eb7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.535648] env[61006]: DEBUG oslo_vmware.rw_handles [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52346f3a-fb47-b0fc-ba48-f63b1734fd46/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 896.535854] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploaded image e72dabda-ebc9-4e91-a9b2-f12e7a212e79 to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 896.540773] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 896.540773] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-108b7a7f-a542-480d-be5a-e63ebf19b946 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.552404] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 896.552404] env[61006]: value = "task-1337191" [ 896.552404] env[61006]: _type = "Task" [ 896.552404] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.562134] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337191, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.579948] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337190, 'name': PowerOffVM_Task, 'duration_secs': 0.216547} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.580254] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.580437] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 17 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 896.719810] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Successfully updated port: 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.758662] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574da64f-0595-4df3-b156-c93005324947 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.766523] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cc801e-6d35-4572-b7cc-1ff5e599098a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.801118] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967beff4-0e3e-4779-a9db-90e1d225fd93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.813834] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bb9df5-3061-47e5-997e-ca05edbcad5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.833966] env[61006]: DEBUG nova.compute.provider_tree [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.835196] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563516} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.839019] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] c5d6f6df-d296-4d4d-815c-159dad2d08dc/c5d6f6df-d296-4d4d-815c-159dad2d08dc.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.839019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.839019] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08c5cd04-ff88-4106-9c67-77436614fe6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.841978] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 896.841978] env[61006]: value = "task-1337192" [ 896.841978] env[61006]: _type = "Task" [ 896.841978] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.850073] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.940145] env[61006]: DEBUG nova.network.neutron [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.946538] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.974607] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31c1433e-5cf8-42fe-8910-e627cbf8131a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.982838] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff248f8-1797-44c4-a4cf-851a515738e8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.016144] env[61006]: DEBUG nova.compute.manager [req-382ec1e6-ad7a-469f-8be1-4d032afaa2ba req-00bdd7e6-b017-4915-ac55-e3d86cbc8474 service nova] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Detach interface failed, port_id=0c39072b-7039-4a65-98e8-728f980a06b6, reason: Instance 84befce1-ccd4-4937-9656-e5cbb7882f47 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 897.017285] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Successfully created port: 7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.062054] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337191, 'name': Destroy_Task, 'duration_secs': 0.305772} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.062468] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroyed the VM [ 897.062934] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 897.063381] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-175bd4d9-96c7-46fa-a197-bbe9d134bf0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.069411] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 897.069411] env[61006]: value = "task-1337193" [ 897.069411] env[61006]: _type = "Task" [ 897.069411] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.077625] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337193, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.089479] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.090267] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.090599] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.090917] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.091211] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.091491] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.091913] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.092344] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.092808] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.093271] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.093732] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.106021] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-481a2b27-5e3e-4f07-bbc4-f1ec469276f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.123434] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 897.123434] env[61006]: value = "task-1337194" [ 897.123434] env[61006]: _type = "Task" [ 897.123434] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.134116] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337194, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.224280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.224280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.224280] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 897.341203] env[61006]: DEBUG nova.scheduler.client.report [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.358022] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070906} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.358022] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.358022] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8664c906-3c2e-42a2-8858-3e28336abcbb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.383259] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] c5d6f6df-d296-4d4d-815c-159dad2d08dc/c5d6f6df-d296-4d4d-815c-159dad2d08dc.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.384407] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e138b1d-313d-4477-aab9-6217a67915a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.405249] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 897.405249] env[61006]: value = "task-1337195" [ 897.405249] env[61006]: _type = "Task" [ 897.405249] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.415161] env[61006]: DEBUG nova.network.neutron [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.416437] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337195, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.442852] env[61006]: INFO nova.compute.manager [-] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Took 1.26 seconds to deallocate network for instance. [ 897.579231] env[61006]: DEBUG oslo_vmware.api [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337193, 'name': RemoveSnapshot_Task, 'duration_secs': 0.493197} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.579516] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 897.579760] env[61006]: INFO nova.compute.manager [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 16.95 seconds to snapshot the instance on the hypervisor. [ 897.636051] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337194, 'name': ReconfigVM_Task, 'duration_secs': 0.34591} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.636379] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 33 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 897.758175] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.850813] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.853809] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.172s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.853922] env[61006]: DEBUG nova.objects.instance [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'resources' on Instance uuid cca6374c-09a4-4145-a116-c49c5a8330c1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.885744] env[61006]: INFO nova.scheduler.client.report [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e [ 897.919596] env[61006]: INFO nova.compute.manager [-] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Took 1.46 seconds to deallocate network for instance. [ 897.920041] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.950427] env[61006]: DEBUG nova.network.neutron [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.952322] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.956654] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.990870] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.991157] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.991317] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.991498] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.991640] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.991793] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.992126] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.993559] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.993559] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.993559] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.993559] env[61006]: DEBUG nova.virt.hardware [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.995747] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca6826b-b087-4965-ab1a-77ae1b598b55 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.003799] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e4e9fc-a1ed-40d7-8eea-f4a28331b293 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.144519] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.144930] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.145174] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.145368] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.145518] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.145680] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.146102] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.146340] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.147503] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.147503] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.147503] env[61006]: DEBUG nova.virt.hardware [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.153774] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 898.156663] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90589161-c0e7-493f-8868-2a7d1f2339a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.178442] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 898.178442] env[61006]: value = "task-1337196" [ 898.178442] env[61006]: _type = "Task" [ 898.178442] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.183794] env[61006]: DEBUG nova.compute.manager [None req-46282151-56d7-460f-8ab7-b9eba4fe9e07 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Found 1 images (rotation: 2) {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 898.190965] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337196, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.394298] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57cb5985-102a-4ecd-8c07-0ab8d1959f27 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.999s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.419420] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337195, 'name': ReconfigVM_Task, 'duration_secs': 0.936879} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.420285] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Reconfigured VM instance instance-00000054 to attach disk [datastore2] c5d6f6df-d296-4d4d-815c-159dad2d08dc/c5d6f6df-d296-4d4d-815c-159dad2d08dc.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.420629] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16dd79d5-e341-4140-b4a7-801cb749b717 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.428606] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.429168] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 898.429168] env[61006]: value = "task-1337197" [ 898.429168] env[61006]: _type = "Task" [ 898.429168] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.444342] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337197, 'name': Rename_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.453370] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.453722] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Instance network_info: |[{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 898.454244] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:54:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43c495ff-c9d9-46fd-93e5-5c92d5f01d1f', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.461872] env[61006]: DEBUG oslo.service.loopingcall [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.464556] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 898.464994] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2b9672f-c050-4f0f-9876-aadb8771966b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.489613] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.489613] env[61006]: value = "task-1337198" [ 898.489613] env[61006]: _type = "Task" [ 898.489613] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.501757] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337198, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.377123] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Successfully updated port: 7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.381719] env[61006]: DEBUG nova.compute.manager [req-15ff00af-f4af-4252-8cb2-acd4c900740e req-44871bdc-1dde-4f7d-9382-7ac48d6ae7da service nova] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Received event network-vif-deleted-a7170e1b-5bc7-4be2-a529-25722de06d50 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.382810] env[61006]: DEBUG nova.compute.manager [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.383022] env[61006]: DEBUG nova.compute.manager [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing instance network info cache due to event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.383327] env[61006]: DEBUG oslo_concurrency.lockutils [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.383502] env[61006]: DEBUG oslo_concurrency.lockutils [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.383675] env[61006]: DEBUG nova.network.neutron [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.401494] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337197, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.408170] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337198, 'name': CreateVM_Task, 'duration_secs': 0.312521} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.409118] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61949248-231e-44c8-b61e-7896d25ee1fa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.411852] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337196, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.411978] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 899.414830] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.414830] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.414830] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.414830] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc819f74-4445-49a5-929c-d22867ffba29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.419038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098f340a-6cf3-4813-9152-4570874bd19b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.423027] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 899.423027] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52009085-a27e-7747-f113-b6fa991dd0c0" [ 899.423027] env[61006]: _type = "Task" [ 899.423027] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.451197] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae53d5c-3cfb-46ec-a38f-4f33cfc647c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.457484] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52009085-a27e-7747-f113-b6fa991dd0c0, 'name': SearchDatastore_Task, 'duration_secs': 0.011673} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.458228] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.458475] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.458733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.458881] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.459214] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.459392] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7303a51-e8d7-46ae-8fa0-29724bbf843f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.465096] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2881d2bc-1a70-4e3a-b83f-8260822356e1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.470103] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.470320] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 899.471450] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6da23c14-e6dc-4bc8-86c7-8ab96e20c08f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.483363] env[61006]: DEBUG nova.compute.provider_tree [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.487915] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 899.487915] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1c36b-a276-23f1-632d-e3e10c86a47a" [ 899.487915] env[61006]: _type = "Task" [ 899.487915] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.496143] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1c36b-a276-23f1-632d-e3e10c86a47a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.608859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.609142] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.888120] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.888254] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquired lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.888405] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.889652] env[61006]: DEBUG nova.compute.manager [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.894256] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ca997b-2a1f-4ec2-97dd-22487a5f984a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.897290] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337196, 'name': ReconfigVM_Task, 'duration_secs': 1.252221} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.899833] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 899.904037] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c533ad-31a1-4ded-b2f8-ca1cf74e2a56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.906622] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337197, 'name': Rename_Task, 'duration_secs': 1.123176} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.909496] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.914017] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0efdf1e-b506-4e86-b0bf-0d8fd057e720 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.931597] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.933177] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-673e6612-5d2b-4f53-99ca-6a737bb1265e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.948167] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 899.948167] env[61006]: value = "task-1337199" [ 899.948167] env[61006]: _type = "Task" [ 899.948167] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.959538] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 899.959538] env[61006]: value = "task-1337200" [ 899.959538] env[61006]: _type = "Task" [ 899.959538] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.965742] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337199, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.971691] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337200, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.990484] env[61006]: DEBUG nova.scheduler.client.report [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.006035] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52d1c36b-a276-23f1-632d-e3e10c86a47a, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.007454] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-884444f8-05f8-4f5e-8171-03d16b0063f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.013439] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 900.013439] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f39dd1-dc29-1ab6-084f-794cc331d3d6" [ 900.013439] env[61006]: _type = "Task" [ 900.013439] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.023917] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f39dd1-dc29-1ab6-084f-794cc331d3d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.048677] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "2259c42a-79f8-455f-b29f-464705526932" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.049427] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.112142] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.206828] env[61006]: DEBUG nova.network.neutron [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updated VIF entry in instance network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.207243] env[61006]: DEBUG nova.network.neutron [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.414999] env[61006]: INFO nova.compute.manager [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] instance snapshotting [ 900.415655] env[61006]: DEBUG nova.objects.instance [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.424174] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.461057] env[61006]: DEBUG oslo_vmware.api [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337199, 'name': PowerOnVM_Task, 'duration_secs': 0.474234} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.464194] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.464420] env[61006]: INFO nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Took 9.36 seconds to spawn the instance on the hypervisor. [ 900.464606] env[61006]: DEBUG nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.465667] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca5f047-f74a-40c0-a110-36025d220112 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.472978] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337200, 'name': ReconfigVM_Task, 'duration_secs': 0.334422} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.474475] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 370f2153-adca-4513-8549-2bb7499cf913/370f2153-adca-4513-8549-2bb7499cf913.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.474779] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 50 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 900.502222] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.649s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.504470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.903s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.504707] env[61006]: DEBUG nova.objects.instance [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lazy-loading 'resources' on Instance uuid 2232c77d-9f4e-4981-9f55-614bba5d71db {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.526192] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f39dd1-dc29-1ab6-084f-794cc331d3d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009826} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.526388] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.526620] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/1a546f17-2fb8-4b99-9001-98cc6fe76837.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.527638] env[61006]: INFO nova.scheduler.client.report [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted allocations for instance cca6374c-09a4-4145-a116-c49c5a8330c1 [ 900.531460] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9e664ea-c4ef-4b64-9534-18d9e9151d91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.542775] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 900.542775] env[61006]: value = "task-1337201" [ 900.542775] env[61006]: _type = "Task" [ 900.542775] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.552275] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.556030] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337201, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.638452] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.642183] env[61006]: DEBUG nova.network.neutron [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Updating instance_info_cache with network_info: [{"id": "7ff4563e-2a36-4112-a195-52151bb79f12", "address": "fa:16:3e:ed:85:cb", "network": {"id": "989a7e0d-7413-4cab-86ec-ae7a5125047e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-451812434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e28ff2841a84b969a83d6b63f621eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ff4563e-2a", "ovs_interfaceid": "7ff4563e-2a36-4112-a195-52151bb79f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.683533] env[61006]: DEBUG nova.compute.manager [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Received event network-vif-plugged-7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.683834] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Acquiring lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.684266] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.684527] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.684768] env[61006]: DEBUG nova.compute.manager [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] No waiting events found dispatching network-vif-plugged-7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.685013] env[61006]: WARNING nova.compute.manager [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Received unexpected event network-vif-plugged-7ff4563e-2a36-4112-a195-52151bb79f12 for instance with vm_state building and task_state spawning. [ 900.685256] env[61006]: DEBUG nova.compute.manager [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Received event network-changed-7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.685581] env[61006]: DEBUG nova.compute.manager [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Refreshing instance network info cache due to event network-changed-7ff4563e-2a36-4112-a195-52151bb79f12. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.685706] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Acquiring lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.709833] env[61006]: DEBUG oslo_concurrency.lockutils [req-9952f682-6a2c-4102-9e7a-08aebf671150 req-ad6cf8eb-b0a4-4a87-8f69-f1440098e334 service nova] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.921601] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c3b051-be69-4fe0-88ed-5be4816b78c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.943156] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a28e82a-80c2-46eb-bdde-474847e0a468 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.989580] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc863ba9-2d9b-433f-a296-26c36476c310 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.994922] env[61006]: INFO nova.compute.manager [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Took 30.10 seconds to build instance. [ 901.016558] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fbf934-3bf8-4109-9f1b-07ba716e163a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.037346] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 67 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 901.044016] env[61006]: DEBUG oslo_concurrency.lockutils [None req-97a42795-1652-47b9-9c81-1aa796821531 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "cca6374c-09a4-4145-a116-c49c5a8330c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.596s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.044901] env[61006]: DEBUG oslo_concurrency.lockutils [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] Acquired lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.049865] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f6c23b-2b14-4966-ac1c-8af5ccd31cc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.059818] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337201, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.061254] env[61006]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 901.061429] env[61006]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61006) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 901.066170] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97939570-a129-4a1f-a972-d16699be67fb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.075443] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af0621f-8feb-4c64-953a-478df013c362 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.089174] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.107543] env[61006]: ERROR root [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-285299' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-285299' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-285299' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-285299'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-285299' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-285299' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-285299'}\n"]: nova.exception.InstanceNotFound: Instance cca6374c-09a4-4145-a116-c49c5a8330c1 could not be found. [ 901.107746] env[61006]: DEBUG oslo_concurrency.lockutils [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] Releasing lock "cca6374c-09a4-4145-a116-c49c5a8330c1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.108011] env[61006]: DEBUG nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Detach interface failed, port_id=0df9528c-287c-45f7-8d8d-b43dc00c0b24, reason: Instance cca6374c-09a4-4145-a116-c49c5a8330c1 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 901.108148] env[61006]: DEBUG nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Received event network-vif-deleted-385eff2a-3efb-49b9-8063-c01a39357798 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.108343] env[61006]: INFO nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Neutron deleted interface 385eff2a-3efb-49b9-8063-c01a39357798; detaching it from the instance and deleting it from the info cache [ 901.108571] env[61006]: DEBUG nova.network.neutron [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Updating instance_info_cache with network_info: [{"id": "598c1248-f434-4f05-ac91-c7f1322879fa", "address": "fa:16:3e:9e:67:e3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap598c1248-f4", "ovs_interfaceid": "598c1248-f434-4f05-ac91-c7f1322879fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.145148] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Releasing lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.145460] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Instance network_info: |[{"id": "7ff4563e-2a36-4112-a195-52151bb79f12", "address": "fa:16:3e:ed:85:cb", "network": {"id": "989a7e0d-7413-4cab-86ec-ae7a5125047e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-451812434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e28ff2841a84b969a83d6b63f621eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ff4563e-2a", "ovs_interfaceid": "7ff4563e-2a36-4112-a195-52151bb79f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.145766] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Acquired lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.145946] env[61006]: DEBUG nova.network.neutron [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Refreshing network info cache for port 7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.147349] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:85:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '72781990-3cb3-42eb-9eb1-4040dedbf66f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ff4563e-2a36-4112-a195-52151bb79f12', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.154336] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Creating folder: Project (8e28ff2841a84b969a83d6b63f621eb7). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.157160] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b891954-7d96-436d-baba-ce16c8f0fa87 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.170083] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Created folder: Project (8e28ff2841a84b969a83d6b63f621eb7) in parent group-v285275. [ 901.170212] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Creating folder: Instances. Parent ref: group-v285369. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.172700] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e072799f-5949-4228-ad84-e9d3fd415be3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.182648] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Created folder: Instances in parent group-v285369. [ 901.182648] env[61006]: DEBUG oslo.service.loopingcall [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.182835] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.183075] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e71e6c6-a2f8-4cbb-843e-231d5df87608 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.204318] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.204318] env[61006]: value = "task-1337204" [ 901.204318] env[61006]: _type = "Task" [ 901.204318] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.213800] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337204, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.306447] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4d0662-790d-4e34-90a3-99f9f83f8a61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.314850] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89217b72-ddf3-4b1c-bc7d-2a109fe85ef3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.344064] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d65582b-44ab-4fd8-8bf2-26603986f282 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.351601] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282963b9-6d86-4802-88b5-26bdf84d34bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.365709] env[61006]: DEBUG nova.compute.provider_tree [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.456834] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 901.456834] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cadeaec3-d79a-417c-97a1-d6ec2937d1de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.469023] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 901.469023] env[61006]: value = "task-1337205" [ 901.469023] env[61006]: _type = "Task" [ 901.469023] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.490959] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337205, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.497042] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9ae991ed-8b68-4c1c-ae1d-33214737251b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.613s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.562284] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337201, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524966} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.562284] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/1a546f17-2fb8-4b99-9001-98cc6fe76837.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.562284] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.562560] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac2b4c6e-d39c-4e19-99c0-b4a6fe2e3c64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.571142] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 901.571142] env[61006]: value = "task-1337206" [ 901.571142] env[61006]: _type = "Task" [ 901.571142] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.580505] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.606514] env[61006]: DEBUG nova.network.neutron [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Port f537f6f0-2a6c-4479-84ea-686716df3690 binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 901.613050] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4c7aee5-f501-4741-b53a-14287a2472f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.626979] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d60277-cf3e-4330-9469-84ce6af7bb6e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.664350] env[61006]: DEBUG nova.compute.manager [req-7dcf3a47-56e4-40f8-bcd9-64f6c8331c41 req-8bcb47d9-0d95-4489-b255-f7fd1b4f4cef service nova] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Detach interface failed, port_id=385eff2a-3efb-49b9-8063-c01a39357798, reason: Instance cca6374c-09a4-4145-a116-c49c5a8330c1 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 901.714928] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337204, 'name': CreateVM_Task, 'duration_secs': 0.414135} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.715169] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.715909] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.716101] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.716488] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.716773] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8afef65c-cf78-4e3a-be76-fdb680c686cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.721813] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 901.721813] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e4403e-58ba-a50c-4d4a-66b052a2c60a" [ 901.721813] env[61006]: _type = "Task" [ 901.721813] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.730901] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e4403e-58ba-a50c-4d4a-66b052a2c60a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.868982] env[61006]: DEBUG nova.scheduler.client.report [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.911692] env[61006]: DEBUG nova.network.neutron [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Updated VIF entry in instance network info cache for port 7ff4563e-2a36-4112-a195-52151bb79f12. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.912172] env[61006]: DEBUG nova.network.neutron [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Updating instance_info_cache with network_info: [{"id": "7ff4563e-2a36-4112-a195-52151bb79f12", "address": "fa:16:3e:ed:85:cb", "network": {"id": "989a7e0d-7413-4cab-86ec-ae7a5125047e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-451812434-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e28ff2841a84b969a83d6b63f621eb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ff4563e-2a", "ovs_interfaceid": "7ff4563e-2a36-4112-a195-52151bb79f12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.964958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.965281] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.965561] env[61006]: INFO nova.compute.manager [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Rebooting instance [ 901.979710] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337205, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.079712] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151147} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.080030] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.080784] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997bf821-b1c9-40e4-915b-a2a698dc13a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.103777] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/1a546f17-2fb8-4b99-9001-98cc6fe76837.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.103875] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1edc3358-1b0b-4b5b-81f6-176345ab58df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.131022] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 902.131022] env[61006]: value = "task-1337207" [ 902.131022] env[61006]: _type = "Task" [ 902.131022] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.142022] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337207, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.235449] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e4403e-58ba-a50c-4d4a-66b052a2c60a, 'name': SearchDatastore_Task, 'duration_secs': 0.009376} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.235816] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.236099] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.236375] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.236507] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.236705] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.236994] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea705eb5-5950-4905-84ee-f125cfcc95a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.248344] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.248344] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.248344] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c66eba0a-6730-43bf-b327-503921a4fa53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.254484] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 902.254484] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521ebfd1-fc79-1ab9-be22-9936fc8b83fd" [ 902.254484] env[61006]: _type = "Task" [ 902.254484] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.263243] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521ebfd1-fc79-1ab9-be22-9936fc8b83fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.373795] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.376163] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.922s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.376382] env[61006]: DEBUG nova.objects.instance [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid d1e09ace-177d-430a-9e65-d01368815272 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.396124] env[61006]: INFO nova.scheduler.client.report [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted allocations for instance 2232c77d-9f4e-4981-9f55-614bba5d71db [ 902.418603] env[61006]: DEBUG oslo_concurrency.lockutils [req-6a435794-ab43-437b-8c2f-7c9fc91714bc req-2ec67617-8bb9-423b-8cb0-dd149a9ac5eb service nova] Releasing lock "refresh_cache-b27fc628-1ac8-4283-bf6e-bcff1cbfe149" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.480776] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337205, 'name': CreateSnapshot_Task, 'duration_secs': 0.533562} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.481574] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 902.482938] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.483123] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquired lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.483373] env[61006]: DEBUG nova.network.neutron [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.485222] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd4c856-1860-4a24-97a2-88c0bffb8c94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.643726] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.643971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.644923] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.651112] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337207, 'name': ReconfigVM_Task, 'duration_secs': 0.44987} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.651112] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/1a546f17-2fb8-4b99-9001-98cc6fe76837.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.651695] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad71f993-6d68-40ff-bae7-144547e9e7ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.659058] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 902.659058] env[61006]: value = "task-1337208" [ 902.659058] env[61006]: _type = "Task" [ 902.659058] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.668416] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337208, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.767693] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521ebfd1-fc79-1ab9-be22-9936fc8b83fd, 'name': SearchDatastore_Task, 'duration_secs': 0.023549} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.768592] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7df23630-5d3a-4960-ae7f-333113e993c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.774184] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 902.774184] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528f93ad-b815-fe55-521f-2153056930cb" [ 902.774184] env[61006]: _type = "Task" [ 902.774184] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.782529] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528f93ad-b815-fe55-521f-2153056930cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.906898] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cd8215d5-273f-4cc6-9238-6de08d8e864e tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "2232c77d-9f4e-4981-9f55-614bba5d71db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.187s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.004103] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 903.006897] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-01270e1f-16f1-401a-a3ff-a2b8016d9fc9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.015525] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 903.015525] env[61006]: value = "task-1337209" [ 903.015525] env[61006]: _type = "Task" [ 903.015525] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.027841] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337209, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.124930] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb7af9b-d282-493c-9f17-a1cd4826f3f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.136168] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8128791c-eb2f-41e1-814d-305e74c79c2d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.179023] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf0adea-5e4f-4fa0-a65a-e1ef97aed70c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.186802] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.187065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.196090] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337208, 'name': Rename_Task, 'duration_secs': 0.266727} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.196090] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14984d2-fc3e-46fa-ba50-4bd46076faa9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.199792] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.200803] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a88e342-c8f2-4617-bec6-92f380a08c3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.215210] env[61006]: DEBUG nova.compute.provider_tree [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.218305] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 903.218305] env[61006]: value = "task-1337210" [ 903.218305] env[61006]: _type = "Task" [ 903.218305] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.230648] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337210, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.284423] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528f93ad-b815-fe55-521f-2153056930cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009559} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.284752] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.285035] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] b27fc628-1ac8-4283-bf6e-bcff1cbfe149/b27fc628-1ac8-4283-bf6e-bcff1cbfe149.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.285288] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6197964-f4b9-4b34-8125-d444aff3e1f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.291902] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 903.291902] env[61006]: value = "task-1337211" [ 903.291902] env[61006]: _type = "Task" [ 903.291902] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.300756] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.429794] env[61006]: DEBUG nova.network.neutron [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updating instance_info_cache with network_info: [{"id": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "address": "fa:16:3e:a9:6f:95", "network": {"id": "6bbe0858-571f-4b0c-8912-063ff6b5469c", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-740305265-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3ab42ad314646e792f4ad39bfc40cf8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c66a277b-e3bf-43b8-a632-04fdd0720b91", "external-id": "nsx-vlan-transportzone-665", "segmentation_id": 665, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f257dd0-56", "ovs_interfaceid": "2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.526585] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337209, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.686573] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.686847] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.687108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.687303] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.687498] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.689788] env[61006]: INFO nova.compute.manager [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Terminating instance [ 903.691962] env[61006]: DEBUG nova.compute.manager [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.692209] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.693183] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6470bd-cbc3-4afa-aa21-e3e9da60447d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.697556] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.705377] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.705594] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cb2b310-9305-4405-94ad-9814355adc5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.713027] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 903.713027] env[61006]: value = "task-1337212" [ 903.713027] env[61006]: _type = "Task" [ 903.713027] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.719711] env[61006]: DEBUG nova.scheduler.client.report [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.730700] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.737410] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337210, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.767108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.767335] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.767856] env[61006]: DEBUG nova.network.neutron [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.802477] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337211, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.933071] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Releasing lock "refresh_cache-c5d6f6df-d296-4d4d-815c-159dad2d08dc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.935098] env[61006]: DEBUG nova.compute.manager [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.935992] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9957e3-aa22-4a7c-899b-58994aa4cef1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.027155] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337209, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.171914] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "40500a0d-33f3-492a-9573-ed10e45642c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.172252] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.217812] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.224487] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337212, 'name': PowerOffVM_Task, 'duration_secs': 0.287201} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.227633] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.227851] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 904.228145] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea7dc788-ebd1-431d-8b06-2eb65e0ae973 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.235624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.859s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.237773] env[61006]: DEBUG oslo_vmware.api [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337210, 'name': PowerOnVM_Task, 'duration_secs': 0.805648} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.238251] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.286s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.238787] env[61006]: DEBUG nova.objects.instance [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lazy-loading 'resources' on Instance uuid 84befce1-ccd4-4937-9656-e5cbb7882f47 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.240478] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.240975] env[61006]: INFO nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Took 8.76 seconds to spawn the instance on the hypervisor. [ 904.241180] env[61006]: DEBUG nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.242691] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa2eb87-8338-4edb-9b2b-f148e9e5ab99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.258058] env[61006]: INFO nova.scheduler.client.report [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance d1e09ace-177d-430a-9e65-d01368815272 [ 904.303730] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513433} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.306049] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] b27fc628-1ac8-4283-bf6e-bcff1cbfe149/b27fc628-1ac8-4283-bf6e-bcff1cbfe149.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.306294] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.307793] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8eafa93-d266-42ec-9ba9-5f01da34d456 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.310271] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 904.310420] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 904.310593] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleting the datastore file [datastore2] e2a40bd9-fb66-40a2-bcf1-5c74707d59dd {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.311034] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1926de8-f672-46ba-8302-51f185b185c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.317536] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 904.317536] env[61006]: value = "task-1337215" [ 904.317536] env[61006]: _type = "Task" [ 904.317536] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.318809] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 904.318809] env[61006]: value = "task-1337214" [ 904.318809] env[61006]: _type = "Task" [ 904.318809] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.329368] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337215, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.332384] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.482610] env[61006]: DEBUG nova.network.neutron [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.527242] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337209, 'name': CloneVM_Task, 'duration_secs': 1.339093} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.527546] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created linked-clone VM from snapshot [ 904.528344] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13752cad-34cf-497f-9830-7bbb852e8adc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.536457] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploading image 67e15c09-2637-49df-a3f0-1a7e302e4ceb {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 904.562713] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.562713] env[61006]: value = "vm-285373" [ 904.562713] env[61006]: _type = "VirtualMachine" [ 904.562713] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.563179] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c7dbc02b-6f70-45db-a77b-a328ea4e7d95 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.570363] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease: (returnval){ [ 904.570363] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b38b19-60e3-7085-3d22-c155425bae88" [ 904.570363] env[61006]: _type = "HttpNfcLease" [ 904.570363] env[61006]: } obtained for exporting VM: (result){ [ 904.570363] env[61006]: value = "vm-285373" [ 904.570363] env[61006]: _type = "VirtualMachine" [ 904.570363] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.570626] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the lease: (returnval){ [ 904.570626] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b38b19-60e3-7085-3d22-c155425bae88" [ 904.570626] env[61006]: _type = "HttpNfcLease" [ 904.570626] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.576925] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.576925] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b38b19-60e3-7085-3d22-c155425bae88" [ 904.576925] env[61006]: _type = "HttpNfcLease" [ 904.576925] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.674612] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.769329] env[61006]: INFO nova.compute.manager [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Took 25.36 seconds to build instance. [ 904.770367] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4ddac8b2-6dd0-4033-95f9-61ab7118403c tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d1e09ace-177d-430a-9e65-d01368815272" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.337s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.834423] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337215, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063978} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.837381] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.837751] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.840669] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18506ad8-a937-48d2-a2e8-0ae930fe8f5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.863081] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] b27fc628-1ac8-4283-bf6e-bcff1cbfe149/b27fc628-1ac8-4283-bf6e-bcff1cbfe149.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.866313] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-490c4d92-b145-42a9-a6b6-865e8e55ec84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.886526] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 904.886526] env[61006]: value = "task-1337217" [ 904.886526] env[61006]: _type = "Task" [ 904.886526] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.896104] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337217, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.958339] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f22af2-7745-42b0-abf4-4ccaf029c52b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.965816] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Doing hard reboot of VM {{(pid=61006) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 904.966151] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-cb9cba14-2a59-4f55-ae85-cb9670f1c2b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.974022] env[61006]: DEBUG oslo_vmware.api [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 904.974022] env[61006]: value = "task-1337218" [ 904.974022] env[61006]: _type = "Task" [ 904.974022] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.981300] env[61006]: DEBUG oslo_vmware.api [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337218, 'name': ResetVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.984417] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.036865] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6b63b3-f2d6-43c0-9124-885fcd5b470a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.045850] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b336ee1-8af3-419d-9f36-ba46cba81e1c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.080315] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4a8ea1-dc12-44b0-b542-01f3c9b8fcb4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.086193] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.086193] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b38b19-60e3-7085-3d22-c155425bae88" [ 905.086193] env[61006]: _type = "HttpNfcLease" [ 905.086193] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 905.088197] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 905.088197] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b38b19-60e3-7085-3d22-c155425bae88" [ 905.088197] env[61006]: _type = "HttpNfcLease" [ 905.088197] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 905.088926] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d60fca2-7c44-469e-9da0-c1859f7c0545 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.092120] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5830b89c-3035-4ea4-811b-aa9e85d3b225 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.101337] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 905.101512] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 905.111023] env[61006]: DEBUG nova.compute.provider_tree [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.193253] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.202907] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c0ccf9b5-cb34-4ca0-a699-75c36a62c0a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.272313] env[61006]: DEBUG oslo_concurrency.lockutils [None req-aed5d88f-0b7b-4070-81fb-a667433f80e9 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.874s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.335028] env[61006]: DEBUG oslo_vmware.api [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.625012} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.335028] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.335298] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.335298] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.335463] env[61006]: INFO nova.compute.manager [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Took 1.64 seconds to destroy the instance on the hypervisor. [ 905.335891] env[61006]: DEBUG oslo.service.loopingcall [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.335988] env[61006]: DEBUG nova.compute.manager [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.335988] env[61006]: DEBUG nova.network.neutron [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.397283] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337217, 'name': ReconfigVM_Task, 'duration_secs': 0.27305} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.397775] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Reconfigured VM instance instance-00000056 to attach disk [datastore2] b27fc628-1ac8-4283-bf6e-bcff1cbfe149/b27fc628-1ac8-4283-bf6e-bcff1cbfe149.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.398247] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c79b0727-07f3-4d43-b94f-0aadbfa236bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.405059] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 905.405059] env[61006]: value = "task-1337219" [ 905.405059] env[61006]: _type = "Task" [ 905.405059] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.413817] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337219, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.484055] env[61006]: DEBUG oslo_vmware.api [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337218, 'name': ResetVM_Task, 'duration_secs': 0.103244} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.484747] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Did hard reboot of VM {{(pid=61006) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 905.485178] env[61006]: DEBUG nova.compute.manager [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.486268] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b48b6b-9202-4f8b-8bfc-368a4a11a033 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.512179] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.512631] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.519205] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdbb60a-d50f-4149-b963-541b1552399f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.551219] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e9f99e-b914-4803-b32d-8e3701a6500c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.560748] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 83 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 905.615671] env[61006]: DEBUG nova.scheduler.client.report [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.706533] env[61006]: DEBUG nova.compute.manager [req-1bea6c33-770d-4dca-bae0-07d70f0bfb56 req-98ba69a5-5731-44b5-8def-f71409a3bb53 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Received event network-vif-deleted-fcaabd7b-dda4-4f87-abae-d9d5335eee51 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.706745] env[61006]: INFO nova.compute.manager [req-1bea6c33-770d-4dca-bae0-07d70f0bfb56 req-98ba69a5-5731-44b5-8def-f71409a3bb53 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Neutron deleted interface fcaabd7b-dda4-4f87-abae-d9d5335eee51; detaching it from the instance and deleting it from the info cache [ 905.707634] env[61006]: DEBUG nova.network.neutron [req-1bea6c33-770d-4dca-bae0-07d70f0bfb56 req-98ba69a5-5731-44b5-8def-f71409a3bb53 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.917194] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337219, 'name': Rename_Task, 'duration_secs': 0.175299} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.917350] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.918048] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2dbe1ad-1f1a-4cd3-97fd-7858192ad391 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.924202] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 905.924202] env[61006]: value = "task-1337220" [ 905.924202] env[61006]: _type = "Task" [ 905.924202] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.936046] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337220, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.014568] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c1a26365-8b54-4446-8933-f3bd24135d5b tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.049s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.015681] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.068197] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.069032] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f35328e-bd2c-4f06-a079-954cce7dd445 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.078319] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 906.078319] env[61006]: value = "task-1337221" [ 906.078319] env[61006]: _type = "Task" [ 906.078319] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.092069] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.112438] env[61006]: DEBUG nova.compute.manager [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.113263] env[61006]: DEBUG nova.compute.manager [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing instance network info cache due to event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.113653] env[61006]: DEBUG oslo_concurrency.lockutils [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.113940] env[61006]: DEBUG oslo_concurrency.lockutils [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.114174] env[61006]: DEBUG nova.network.neutron [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.121118] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.123555] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.695s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.123810] env[61006]: DEBUG nova.objects.instance [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lazy-loading 'resources' on Instance uuid bdcf0445-6f8b-4e7b-bfb3-93a785162e56 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.156721] env[61006]: INFO nova.scheduler.client.report [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted allocations for instance 84befce1-ccd4-4937-9656-e5cbb7882f47 [ 906.177485] env[61006]: DEBUG nova.network.neutron [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.212971] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a81876e1-68ab-4b60-86f3-3f8c749846fc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.224249] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1f8990-2543-4dcd-9154-c983ce32448c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.263020] env[61006]: DEBUG nova.compute.manager [req-1bea6c33-770d-4dca-bae0-07d70f0bfb56 req-98ba69a5-5731-44b5-8def-f71409a3bb53 service nova] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Detach interface failed, port_id=fcaabd7b-dda4-4f87-abae-d9d5335eee51, reason: Instance e2a40bd9-fb66-40a2-bcf1-5c74707d59dd could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 906.436124] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337220, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.545406] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.588451] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337221, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.672570] env[61006]: DEBUG oslo_concurrency.lockutils [None req-1825872f-39c6-4bbf-a95d-5488de7e8d99 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "84befce1-ccd4-4937-9656-e5cbb7882f47" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.609s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.680568] env[61006]: INFO nova.compute.manager [-] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Took 1.34 seconds to deallocate network for instance. [ 906.763803] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.764110] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.858416] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.858776] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.859118] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.859420] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.859887] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.862564] env[61006]: INFO nova.compute.manager [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Terminating instance [ 906.867063] env[61006]: DEBUG nova.compute.manager [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.867701] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.868716] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b99a23a-bd97-4af0-8f72-9701b840f373 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.880062] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.880391] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31b57c97-5625-44c5-8e6a-698e28a428bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.889607] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 906.889607] env[61006]: value = "task-1337222" [ 906.889607] env[61006]: _type = "Task" [ 906.889607] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.904595] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.931841] env[61006]: DEBUG nova.network.neutron [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updated VIF entry in instance network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.932501] env[61006]: DEBUG nova.network.neutron [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.934436] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0ba578-82a0-45f2-8165-f55493829bf2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.942810] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337220, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.945799] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dea5b48-4f1d-4476-bbfd-fff96a8c5e88 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.979946] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bc09cf-ef74-4a19-acf0-09d18be9ec78 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.987799] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1069a813-1859-4c65-b53c-a1ab66ed41ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.003737] env[61006]: DEBUG nova.compute.provider_tree [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.087717] env[61006]: DEBUG oslo_vmware.api [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337221, 'name': PowerOnVM_Task, 'duration_secs': 0.520968} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.087990] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.088199] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0777e02e-7955-49d9-954f-578fee58e31c tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance '370f2153-adca-4513-8549-2bb7499cf913' progress to 100 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 907.191511] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.270207] env[61006]: DEBUG nova.compute.utils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.403024] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337222, 'name': PowerOffVM_Task, 'duration_secs': 0.263462} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.403320] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.403496] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.403757] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8efb017-0fb0-43cf-a3c5-94664cea175c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.436402] env[61006]: DEBUG oslo_vmware.api [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337220, 'name': PowerOnVM_Task, 'duration_secs': 1.233202} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.436665] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.437012] env[61006]: INFO nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 9.48 seconds to spawn the instance on the hypervisor. [ 907.437260] env[61006]: DEBUG nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.437734] env[61006]: DEBUG oslo_concurrency.lockutils [req-4c3a1f7f-e2cd-4e89-a78e-1510b2caea28 req-6dc25a66-627d-456c-a292-36e4273b14e4 service nova] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.438667] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86b0a0a-e2d5-49af-96b1-4293d398266e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.477075] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.477320] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.477520] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Deleting the datastore file [datastore2] c5d6f6df-d296-4d4d-815c-159dad2d08dc {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.477845] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7544c7f-12b8-4598-b045-631bbf0420cf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.485225] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for the task: (returnval){ [ 907.485225] env[61006]: value = "task-1337224" [ 907.485225] env[61006]: _type = "Task" [ 907.485225] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.493224] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337224, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.506552] env[61006]: DEBUG nova.scheduler.client.report [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.773528] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.956970] env[61006]: INFO nova.compute.manager [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 26.24 seconds to build instance. [ 907.995112] env[61006]: DEBUG oslo_vmware.api [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Task: {'id': task-1337224, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.498888} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.995379] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.995565] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.995746] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.995915] env[61006]: INFO nova.compute.manager [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Took 1.13 seconds to destroy the instance on the hypervisor. [ 907.996189] env[61006]: DEBUG oslo.service.loopingcall [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.996389] env[61006]: DEBUG nova.compute.manager [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.996482] env[61006]: DEBUG nova.network.neutron [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 908.013336] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.015594] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.377s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.017319] env[61006]: INFO nova.compute.claims [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.037946] env[61006]: INFO nova.scheduler.client.report [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Deleted allocations for instance bdcf0445-6f8b-4e7b-bfb3-93a785162e56 [ 908.349623] env[61006]: DEBUG nova.compute.manager [req-ae3373b3-044e-42b5-8180-620b29a41230 req-e78e8004-b7c4-420f-8f33-6b5c42ad7d0c service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Received event network-vif-deleted-2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.349883] env[61006]: INFO nova.compute.manager [req-ae3373b3-044e-42b5-8180-620b29a41230 req-e78e8004-b7c4-420f-8f33-6b5c42ad7d0c service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Neutron deleted interface 2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57; detaching it from the instance and deleting it from the info cache [ 908.350115] env[61006]: DEBUG nova.network.neutron [req-ae3373b3-044e-42b5-8180-620b29a41230 req-e78e8004-b7c4-420f-8f33-6b5c42ad7d0c service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.458914] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ff5a391f-d3d5-402b-9a7a-d5aa0e87309c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.759s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.548837] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6340afe5-3818-4cc4-97f0-1137155c6621 tempest-MultipleCreateTestJSON-1280834283 tempest-MultipleCreateTestJSON-1280834283-project-member] Lock "bdcf0445-6f8b-4e7b-bfb3-93a785162e56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.229s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.748411] env[61006]: DEBUG nova.network.neutron [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.834051] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.834441] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.836734] env[61006]: INFO nova.compute.manager [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attaching volume e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83 to /dev/sdb [ 908.853343] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78b6067b-f050-4237-899f-af5338529487 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.867513] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e671a8-33db-4527-9bc5-7f591c9293ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.883615] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d54eb7-5f1a-44ce-b94d-b5172b564261 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.893204] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caacae05-e1d3-49e4-8f3a-524db096037c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.906779] env[61006]: DEBUG nova.compute.manager [req-ae3373b3-044e-42b5-8180-620b29a41230 req-e78e8004-b7c4-420f-8f33-6b5c42ad7d0c service nova] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Detach interface failed, port_id=2f257dd0-56e9-44c5-8f3b-f9f8a9d39d57, reason: Instance c5d6f6df-d296-4d4d-815c-159dad2d08dc could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 908.912143] env[61006]: DEBUG nova.virt.block_device [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating existing volume attachment record: 8406575e-add1-4a51-bbfd-83fd609b8014 {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 909.237478] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.237769] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.237966] env[61006]: DEBUG nova.compute.manager [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Going to confirm migration 1 {{(pid=61006) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 909.249662] env[61006]: INFO nova.compute.manager [-] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Took 1.25 seconds to deallocate network for instance. [ 909.296157] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b29f46-fd23-4fa0-a9fc-9488109b84b9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.305795] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3b6609-5214-4881-8a90-c8cc31e7d9a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.338719] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1544bc49-bc66-449f-a3b0-e321e26d6e29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.346634] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69c59ee-4849-4228-8e41-dd0ba0dd2fe9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.361478] env[61006]: DEBUG nova.compute.provider_tree [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.758257] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.865706] env[61006]: DEBUG nova.scheduler.client.report [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.870384] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.870584] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.870815] env[61006]: DEBUG nova.network.neutron [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 909.871294] env[61006]: DEBUG nova.objects.instance [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lazy-loading 'info_cache' on Instance uuid 370f2153-adca-4513-8549-2bb7499cf913 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.960431] env[61006]: DEBUG nova.compute.manager [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.961492] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361dc723-01c5-42b9-b7ea-d827a0958748 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.374484] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.375056] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.379463] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.290s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.380755] env[61006]: INFO nova.compute.claims [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.475538] env[61006]: INFO nova.compute.manager [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] instance snapshotting [ 910.479218] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661b38b8-7c71-4f09-af4d-c5ac3514a81b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.499322] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca3420e-ffcd-495b-bdba-6088010bd20d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.881564] env[61006]: DEBUG nova.compute.utils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.882945] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.883183] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 910.972911] env[61006]: DEBUG nova.policy [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '744cc90a7b3d45d48571b5821c111588', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97c41f3b09774d639709be60d7a3edf6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.009827] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 911.010515] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b28d89f7-a94e-455b-a04e-b6915659eed0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.020816] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 911.020816] env[61006]: value = "task-1337228" [ 911.020816] env[61006]: _type = "Task" [ 911.020816] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.030749] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337228, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.266840] env[61006]: DEBUG nova.network.neutron [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [{"id": "f537f6f0-2a6c-4479-84ea-686716df3690", "address": "fa:16:3e:fc:f0:29", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf537f6f0-2a", "ovs_interfaceid": "f537f6f0-2a6c-4479-84ea-686716df3690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.386972] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.421720] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Successfully created port: beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.537063] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337228, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.649148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cb67fe-8d0e-474d-86f1-31347b4830bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.657148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2fabc6-c643-4afd-bfe5-afd50ebe7357 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.689243] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4260cd8-8fbd-4daf-b2aa-ce9c09077d0f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.697114] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd7fc7b-01fc-4897-a65b-0a756333b8bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.711772] env[61006]: DEBUG nova.compute.provider_tree [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.769609] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-370f2153-adca-4513-8549-2bb7499cf913" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.769997] env[61006]: DEBUG nova.objects.instance [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lazy-loading 'migration_context' on Instance uuid 370f2153-adca-4513-8549-2bb7499cf913 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.034519] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337228, 'name': CreateSnapshot_Task, 'duration_secs': 0.983975} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.034519] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 912.037209] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aba4ccd-d139-43fa-a727-fb5414cab2f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.215885] env[61006]: DEBUG nova.scheduler.client.report [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.272796] env[61006]: DEBUG nova.objects.base [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Object Instance<370f2153-adca-4513-8549-2bb7499cf913> lazy-loaded attributes: info_cache,migration_context {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 912.273898] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62196505-9a1c-4b6a-a403-3b1a96b879cf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.293175] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f12df6-4f1b-44d2-8e94-699c1b7f3e98 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.298671] env[61006]: DEBUG oslo_vmware.api [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 912.298671] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f477be-c5b0-aea5-bc86-281ccce81ad4" [ 912.298671] env[61006]: _type = "Task" [ 912.298671] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.306080] env[61006]: DEBUG oslo_vmware.api [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f477be-c5b0-aea5-bc86-281ccce81ad4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.401078] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.421998] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.422311] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.422473] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.422681] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.422832] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.423023] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.423254] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.423437] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.423618] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.423785] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.423959] env[61006]: DEBUG nova.virt.hardware [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.424891] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ca8ea5-e9d9-400b-854d-ee45bd70e4b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.433221] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb1ef0c-82f3-456a-b36c-95f35bc121e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.556249] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 912.556616] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-418db34e-c101-4d3c-85d1-9199262459b6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.565676] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 912.565676] env[61006]: value = "task-1337230" [ 912.565676] env[61006]: _type = "Task" [ 912.565676] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.573745] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337230, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.721249] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.721822] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.724590] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.507s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.726026] env[61006]: INFO nova.compute.claims [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.809093] env[61006]: DEBUG oslo_vmware.api [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f477be-c5b0-aea5-bc86-281ccce81ad4, 'name': SearchDatastore_Task, 'duration_secs': 0.007754} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.809486] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.878204] env[61006]: DEBUG nova.compute.manager [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Received event network-vif-plugged-beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.878521] env[61006]: DEBUG oslo_concurrency.lockutils [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] Acquiring lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.878855] env[61006]: DEBUG oslo_concurrency.lockutils [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.879068] env[61006]: DEBUG oslo_concurrency.lockutils [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.879272] env[61006]: DEBUG nova.compute.manager [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] No waiting events found dispatching network-vif-plugged-beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.879467] env[61006]: WARNING nova.compute.manager [req-c26b95f1-1180-42ef-ac49-7cbb7e034180 req-0cca32ee-9a17-4661-ba75-348bc8781024 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Received unexpected event network-vif-plugged-beb7410d-8ed5-460f-937c-81ace8d6600c for instance with vm_state building and task_state spawning. [ 912.974472] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Successfully updated port: beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.076441] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337230, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.231233] env[61006]: DEBUG nova.compute.utils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.236792] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.236792] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.291300] env[61006]: DEBUG nova.policy [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.466162] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 913.466304] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285375', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'name': 'volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'serial': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 913.467208] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04adb61d-e39b-4b09-b923-47c5e3fee32f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.484427] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.484635] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquired lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.484802] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.487043] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be10d62-69c5-49d2-ba5c-eea7ab9cfc12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.515372] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83/volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.516867] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0959ebd-e401-4a96-b6e1-cf493b1213c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.540262] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 913.540262] env[61006]: value = "task-1337231" [ 913.540262] env[61006]: _type = "Task" [ 913.540262] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.552241] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.577021] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337230, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.599678] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Successfully created port: 919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.737078] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.992516] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61422679-65db-41ab-bea2-dac0f4f31b0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.003508] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d96c7b3-6213-42c0-8de5-c4fff57c39c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.050524] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.057684] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f817e12-684c-48ab-9d70-0e2a01259502 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.067778] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337231, 'name': ReconfigVM_Task, 'duration_secs': 0.455405} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.074337] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83/volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.084009] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55e96da5-b906-4aee-8544-afafbeef120c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.100826] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e1234a-cb8a-4cf1-a181-dc53adcc0a27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.113684] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337230, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.127702] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 914.127702] env[61006]: value = "task-1337232" [ 914.127702] env[61006]: _type = "Task" [ 914.127702] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.128757] env[61006]: DEBUG nova.compute.provider_tree [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.143102] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.251305] env[61006]: DEBUG nova.network.neutron [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Updating instance_info_cache with network_info: [{"id": "beb7410d-8ed5-460f-937c-81ace8d6600c", "address": "fa:16:3e:9f:f0:91", "network": {"id": "6be2a04a-ee74-44f5-92fa-ffdd070c1b43", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-590905073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97c41f3b09774d639709be60d7a3edf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb7410d-8e", "ovs_interfaceid": "beb7410d-8ed5-460f-937c-81ace8d6600c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.580024] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337230, 'name': CloneVM_Task, 'duration_secs': 2.002929} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.580024] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Created linked-clone VM from snapshot [ 914.580443] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c90c37-8b88-49f4-bdd7-f35273f47e07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.588908] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Uploading image 07e07d31-cef3-4ce6-a79e-0a94d33e19b9 {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 914.635219] env[61006]: DEBUG nova.scheduler.client.report [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.641517] env[61006]: DEBUG oslo_vmware.api [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337232, 'name': ReconfigVM_Task, 'duration_secs': 0.152829} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.642092] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285375', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'name': 'volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'serial': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 914.749700] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.754398] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Releasing lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.754708] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Instance network_info: |[{"id": "beb7410d-8ed5-460f-937c-81ace8d6600c", "address": "fa:16:3e:9f:f0:91", "network": {"id": "6be2a04a-ee74-44f5-92fa-ffdd070c1b43", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-590905073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97c41f3b09774d639709be60d7a3edf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb7410d-8e", "ovs_interfaceid": "beb7410d-8ed5-460f-937c-81ace8d6600c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.755143] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:f0:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11032cc2-b275-48d2-9c40-9455ea7d49e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'beb7410d-8ed5-460f-937c-81ace8d6600c', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.763757] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Creating folder: Project (97c41f3b09774d639709be60d7a3edf6). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.764072] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2169d684-80d0-4aea-98f0-f9c22ba554a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.774951] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Created folder: Project (97c41f3b09774d639709be60d7a3edf6) in parent group-v285275. [ 914.775575] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Creating folder: Instances. Parent ref: group-v285378. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.775575] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f28e073f-b9f8-4258-8e34-75dd8160eef4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.784920] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Created folder: Instances in parent group-v285378. [ 914.785172] env[61006]: DEBUG oslo.service.loopingcall [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.785365] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.785566] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71affdf3-ef95-4e7a-9021-6efdcf0bba64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.805360] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.805360] env[61006]: value = "task-1337235" [ 914.805360] env[61006]: _type = "Task" [ 914.805360] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.812981] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337235, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.919100] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 914.919100] env[61006]: value = "vm-285377" [ 914.919100] env[61006]: _type = "VirtualMachine" [ 914.919100] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 914.922515] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4a86cf41-b70a-42e1-9a93-63c69c75c9a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.925900] env[61006]: DEBUG nova.compute.manager [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Received event network-changed-beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.926183] env[61006]: DEBUG nova.compute.manager [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Refreshing instance network info cache due to event network-changed-beb7410d-8ed5-460f-937c-81ace8d6600c. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.926562] env[61006]: DEBUG oslo_concurrency.lockutils [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] Acquiring lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.926764] env[61006]: DEBUG oslo_concurrency.lockutils [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] Acquired lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.926965] env[61006]: DEBUG nova.network.neutron [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Refreshing network info cache for port beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.931886] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.932287] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.932510] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.932803] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.933294] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.933294] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.933474] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.933746] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.934038] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.935035] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.935035] env[61006]: DEBUG nova.virt.hardware [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.936259] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0066a6f1-4c02-42dc-8a03-c0a1bb71f749 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.942518] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lease: (returnval){ [ 914.942518] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52837f2a-0c91-9732-625f-80b48cef4cd2" [ 914.942518] env[61006]: _type = "HttpNfcLease" [ 914.942518] env[61006]: } obtained for exporting VM: (result){ [ 914.942518] env[61006]: value = "vm-285377" [ 914.942518] env[61006]: _type = "VirtualMachine" [ 914.942518] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 914.942910] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the lease: (returnval){ [ 914.942910] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52837f2a-0c91-9732-625f-80b48cef4cd2" [ 914.942910] env[61006]: _type = "HttpNfcLease" [ 914.942910] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 914.950566] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7ec1b7-dff7-42e8-a73a-a42daf3974de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.956760] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.956760] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52837f2a-0c91-9732-625f-80b48cef4cd2" [ 914.956760] env[61006]: _type = "HttpNfcLease" [ 914.956760] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 915.143834] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.144569] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.147652] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.955s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.149547] env[61006]: INFO nova.compute.claims [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.189554] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Successfully updated port: 919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.316646] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337235, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.406032] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.406196] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.406436] env[61006]: INFO nova.compute.manager [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Shelving [ 915.452020] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 915.452020] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52837f2a-0c91-9732-625f-80b48cef4cd2" [ 915.452020] env[61006]: _type = "HttpNfcLease" [ 915.452020] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 915.452869] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 915.452869] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52837f2a-0c91-9732-625f-80b48cef4cd2" [ 915.452869] env[61006]: _type = "HttpNfcLease" [ 915.452869] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 915.453169] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ad83aa-37c7-4f57-8d0c-7d0354ba2149 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.461013] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 915.461255] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 915.590406] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c010ec6f-b5aa-4962-ab85-7eb999581e2f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.656502] env[61006]: DEBUG nova.compute.utils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.659590] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.659761] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.692246] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.692397] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.692550] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.699480] env[61006]: DEBUG nova.objects.instance [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lazy-loading 'flavor' on Instance uuid d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.738041] env[61006]: DEBUG nova.policy [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.789293] env[61006]: DEBUG nova.network.neutron [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Updated VIF entry in instance network info cache for port beb7410d-8ed5-460f-937c-81ace8d6600c. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.789679] env[61006]: DEBUG nova.network.neutron [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Updating instance_info_cache with network_info: [{"id": "beb7410d-8ed5-460f-937c-81ace8d6600c", "address": "fa:16:3e:9f:f0:91", "network": {"id": "6be2a04a-ee74-44f5-92fa-ffdd070c1b43", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-590905073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97c41f3b09774d639709be60d7a3edf6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb7410d-8e", "ovs_interfaceid": "beb7410d-8ed5-460f-937c-81ace8d6600c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.816614] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337235, 'name': CreateVM_Task, 'duration_secs': 0.607495} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.816788] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.817509] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.817686] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.818036] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.818431] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db79c9ec-3eda-49fb-949a-d3c6053137ae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.823607] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 915.823607] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526143b0-e6cc-9991-5f97-5e6943ae8a0c" [ 915.823607] env[61006]: _type = "Task" [ 915.823607] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.832239] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526143b0-e6cc-9991-5f97-5e6943ae8a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.915087] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.915405] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ca2ebbe-d0d2-43fd-8444-3c896d8603f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.924261] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 915.924261] env[61006]: value = "task-1337237" [ 915.924261] env[61006]: _type = "Task" [ 915.924261] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.933338] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337237, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.942994] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.944083] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44e7665-51ae-4982-a2c4-dc56b3a07cb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.950685] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.950771] env[61006]: ERROR oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk due to incomplete transfer. [ 915.951081] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a6045420-9af3-48e2-875c-3d8c4c2a9e75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.963723] env[61006]: DEBUG oslo_vmware.rw_handles [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52716d22-5091-ca8f-d263-6dc90f1a5b64/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 915.963981] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploaded image 67e15c09-2637-49df-a3f0-1a7e302e4ceb to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 915.966581] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.966957] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ea2d28d8-eaf1-455f-8199-c8419debff24 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.972876] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 915.972876] env[61006]: value = "task-1337238" [ 915.972876] env[61006]: _type = "Task" [ 915.972876] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.981823] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337238, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.149444] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Successfully created port: 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.162633] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.207239] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f3ef4d91-3aa8-4b04-9b33-8df9d2de9bc6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.373s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.255496] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 916.295582] env[61006]: DEBUG oslo_concurrency.lockutils [req-7451ee98-6c4b-4f55-a0c0-a42ad5220bd3 req-6ac3d2c4-24b2-4e52-9965-dbbc19e5c460 service nova] Releasing lock "refresh_cache-1d6ede6a-5601-4da4-baa1-5090a112ec2d" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.336250] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526143b0-e6cc-9991-5f97-5e6943ae8a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.012554} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.336845] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.337021] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.337267] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.337419] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.337604] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.340950] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4aae252e-9a23-4009-aed7-187142062e1b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.349962] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.350273] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 916.351280] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-615bf049-067e-4e85-a97a-4495aa988a6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.359694] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 916.359694] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd8b96-034c-82db-c430-6607482d33bf" [ 916.359694] env[61006]: _type = "Task" [ 916.359694] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.368017] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd8b96-034c-82db-c430-6607482d33bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.435272] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337237, 'name': PowerOffVM_Task, 'duration_secs': 0.194505} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.437462] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.438578] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c373e4-bd07-4a86-9cc6-27597fcb49e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.443035] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6805ff9c-7021-434c-82d9-0264c022e040 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.465435] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b2b246-8e17-4b67-bd45-5c058250cd80 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.469188] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717b0051-feda-4c98-99c8-43b7fd8dd2b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.511917] env[61006]: DEBUG nova.network.neutron [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Updating instance_info_cache with network_info: [{"id": "919d334d-6479-4c18-a101-55dab88ea5de", "address": "fa:16:3e:85:a6:f8", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919d334d-64", "ovs_interfaceid": "919d334d-6479-4c18-a101-55dab88ea5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.514402] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a3c451-0c06-4d34-84d3-a65ca956d6c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.523455] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337238, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.527520] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa374773-d245-4b9d-a58f-a242b920c747 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.543531] env[61006]: DEBUG nova.compute.provider_tree [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.870926] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd8b96-034c-82db-c430-6607482d33bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.872132] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c993422-c230-4ab7-b03e-2b8694d76ce5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.877943] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 916.877943] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b9af67-a928-7a05-d8ad-14921547b498" [ 916.877943] env[61006]: _type = "Task" [ 916.877943] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.885693] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b9af67-a928-7a05-d8ad-14921547b498, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.950186] env[61006]: DEBUG nova.compute.manager [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Received event network-vif-plugged-919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.950781] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Acquiring lock "2259c42a-79f8-455f-b29f-464705526932-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.951063] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Lock "2259c42a-79f8-455f-b29f-464705526932-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.951261] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Lock "2259c42a-79f8-455f-b29f-464705526932-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.951460] env[61006]: DEBUG nova.compute.manager [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] No waiting events found dispatching network-vif-plugged-919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.951595] env[61006]: WARNING nova.compute.manager [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Received unexpected event network-vif-plugged-919d334d-6479-4c18-a101-55dab88ea5de for instance with vm_state building and task_state spawning. [ 916.951947] env[61006]: DEBUG nova.compute.manager [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Received event network-changed-919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.951947] env[61006]: DEBUG nova.compute.manager [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Refreshing instance network info cache due to event network-changed-919d334d-6479-4c18-a101-55dab88ea5de. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.952220] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Acquiring lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.983381] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.983625] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.989079] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 916.989655] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337238, 'name': Destroy_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.989960] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ca56c916-3541-4b07-8766-ce953d2acc53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.000746] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 917.000746] env[61006]: value = "task-1337239" [ 917.000746] env[61006]: _type = "Task" [ 917.000746] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.009960] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337239, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.013818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.014130] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Instance network_info: |[{"id": "919d334d-6479-4c18-a101-55dab88ea5de", "address": "fa:16:3e:85:a6:f8", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919d334d-64", "ovs_interfaceid": "919d334d-6479-4c18-a101-55dab88ea5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.014419] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Acquired lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.014597] env[61006]: DEBUG nova.network.neutron [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Refreshing network info cache for port 919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.015747] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:a6:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37333dc2-982e-45e9-9dda-0c18417d7fa6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '919d334d-6479-4c18-a101-55dab88ea5de', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.024072] env[61006]: DEBUG oslo.service.loopingcall [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.026860] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.027520] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6f3d26f-ec2b-4f11-9c29-90b52974a67f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.048292] env[61006]: DEBUG nova.scheduler.client.report [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.051846] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.051846] env[61006]: value = "task-1337240" [ 917.051846] env[61006]: _type = "Task" [ 917.051846] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.062264] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337240, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.173061] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.199451] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.200216] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.200216] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.200216] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.200216] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.200435] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.200571] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.200750] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.200993] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.201196] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.201377] env[61006]: DEBUG nova.virt.hardware [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.202596] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721114b2-2889-4bc8-9f6d-d0045a90f734 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.211041] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4820f3c-4e30-4ded-a1a6-e3aa5de0a552 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.264828] env[61006]: DEBUG nova.network.neutron [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Updated VIF entry in instance network info cache for port 919d334d-6479-4c18-a101-55dab88ea5de. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.265358] env[61006]: DEBUG nova.network.neutron [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Updating instance_info_cache with network_info: [{"id": "919d334d-6479-4c18-a101-55dab88ea5de", "address": "fa:16:3e:85:a6:f8", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919d334d-64", "ovs_interfaceid": "919d334d-6479-4c18-a101-55dab88ea5de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.389234] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b9af67-a928-7a05-d8ad-14921547b498, 'name': SearchDatastore_Task, 'duration_secs': 0.009827} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.389702] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.389702] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1d6ede6a-5601-4da4-baa1-5090a112ec2d/1d6ede6a-5601-4da4-baa1-5090a112ec2d.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.390457] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c11d669c-7993-42a7-aded-b2ef5dc62ed5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.397928] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 917.397928] env[61006]: value = "task-1337241" [ 917.397928] env[61006]: _type = "Task" [ 917.397928] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.407251] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.485505] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337238, 'name': Destroy_Task, 'duration_secs': 1.07728} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.488181] env[61006]: DEBUG nova.compute.utils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.488367] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroyed the VM [ 917.488478] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 917.488948] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-62cb100b-e9d4-4ed2-97f1-83e3b6895dac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.495655] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 917.495655] env[61006]: value = "task-1337242" [ 917.495655] env[61006]: _type = "Task" [ 917.495655] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.508756] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337242, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.513696] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337239, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.553263] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.553905] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.559985] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.015s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.561890] env[61006]: INFO nova.compute.claims [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.574033] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337240, 'name': CreateVM_Task, 'duration_secs': 0.412429} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.575061] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 917.575061] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.575061] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.576026] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.576026] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ad4a6c-a921-42a7-8f2b-01e61fce4229 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.580267] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 917.580267] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52979a3c-a788-4fd5-6ee3-5cc03728f16a" [ 917.580267] env[61006]: _type = "Task" [ 917.580267] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.588496] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52979a3c-a788-4fd5-6ee3-5cc03728f16a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.768023] env[61006]: DEBUG oslo_concurrency.lockutils [req-e207b5b7-cbdf-4e5b-9408-7ed9df10b4da req-9510dd8d-8ce6-4595-8c3b-b0165d3514fa service nova] Releasing lock "refresh_cache-2259c42a-79f8-455f-b29f-464705526932" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.911805] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337241, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.991847] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.009765] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337242, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.014619] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337239, 'name': CreateSnapshot_Task, 'duration_secs': 0.942111} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.018052] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 918.018052] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b9d418-df59-4766-b264-7f4a95b6bb87 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.073052] env[61006]: DEBUG nova.compute.utils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.075134] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.075525] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.093582] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52979a3c-a788-4fd5-6ee3-5cc03728f16a, 'name': SearchDatastore_Task, 'duration_secs': 0.012129} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.093582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.093582] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.093582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.093582] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.093582] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.094157] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8265c4e9-45ca-4225-a524-4b2657fedba9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.103396] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.103849] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.105035] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e63f9d4-6d88-4bfe-b237-f8de2a383cb9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.110317] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 918.110317] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524e8317-3f3b-da9f-242b-5381bccb333c" [ 918.110317] env[61006]: _type = "Task" [ 918.110317] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.118411] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524e8317-3f3b-da9f-242b-5381bccb333c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.167631] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Successfully updated port: 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.187114] env[61006]: DEBUG nova.policy [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a82aea0ad3a245af99868fc53a0b04f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd26132a67fc1412c938a36094a34a433', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.412954] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337241, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581352} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.413493] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1d6ede6a-5601-4da4-baa1-5090a112ec2d/1d6ede6a-5601-4da4-baa1-5090a112ec2d.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.413836] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.414253] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca36368c-3887-459c-aead-7bf5ae7dfab8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.423375] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 918.423375] env[61006]: value = "task-1337243" [ 918.423375] env[61006]: _type = "Task" [ 918.423375] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.435232] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337243, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.456199] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Successfully created port: 7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.508843] env[61006]: DEBUG oslo_vmware.api [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337242, 'name': RemoveSnapshot_Task, 'duration_secs': 0.946612} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.509721] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 918.509721] env[61006]: INFO nova.compute.manager [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 17.59 seconds to snapshot the instance on the hypervisor. [ 918.534638] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 918.535535] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dad2d6a9-68d1-4fbd-8e04-28390c9c44aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.543256] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 918.543256] env[61006]: value = "task-1337244" [ 918.543256] env[61006]: _type = "Task" [ 918.543256] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.554185] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337244, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.580097] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.623463] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524e8317-3f3b-da9f-242b-5381bccb333c, 'name': SearchDatastore_Task, 'duration_secs': 0.012164} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.627608] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d3dd6d2-f7a1-43b3-9370-4b2d617a45ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.633493] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 918.633493] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52913d83-3915-ceb9-e8bd-8ec83c1b973f" [ 918.633493] env[61006]: _type = "Task" [ 918.633493] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.642543] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52913d83-3915-ceb9-e8bd-8ec83c1b973f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.669337] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.669564] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.669623] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.849187] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec72af0-33cb-43fd-b55a-33574603fdfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.857985] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff3924a-791f-4d6c-9c9e-5071eacffc66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.888470] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb411a7a-2548-4982-ab84-f16ef6379f4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.896194] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a0fe50-7d98-4386-ba68-78ec8d67fe4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.911815] env[61006]: DEBUG nova.compute.provider_tree [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.934020] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337243, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119806} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.934344] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.935192] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c237d6b-56c4-492f-aace-9d887c278e22 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.957806] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 1d6ede6a-5601-4da4-baa1-5090a112ec2d/1d6ede6a-5601-4da4-baa1-5090a112ec2d.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.958515] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-986ec7b6-b997-437a-abf7-5b12cb54856e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.978411] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 918.978411] env[61006]: value = "task-1337245" [ 918.978411] env[61006]: _type = "Task" [ 918.978411] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.987251] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.994351] env[61006]: DEBUG nova.compute.manager [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-vif-plugged-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.994586] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.994797] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.995038] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.995143] env[61006]: DEBUG nova.compute.manager [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] No waiting events found dispatching network-vif-plugged-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.995314] env[61006]: WARNING nova.compute.manager [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received unexpected event network-vif-plugged-58bb0db5-4669-4185-8d20-b5a77724df9b for instance with vm_state building and task_state spawning. [ 918.995467] env[61006]: DEBUG nova.compute.manager [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.995643] env[61006]: DEBUG nova.compute.manager [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.995807] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.055697] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337244, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.071392] env[61006]: DEBUG nova.compute.manager [None req-9ccfcd76-1374-42a7-9ebd-e2972f6cbaa9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Found 2 images (rotation: 2) {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 919.093699] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.093984] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.094237] env[61006]: INFO nova.compute.manager [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attaching volume 9f7e0232-99d9-446e-aca8-77c70e411002 to /dev/sdc [ 919.128534] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04955217-45d8-43d6-8f36-a736ac2f35e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.140653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2232e4d0-45cf-43fb-9f2b-b21bd4ac920f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.148244] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52913d83-3915-ceb9-e8bd-8ec83c1b973f, 'name': SearchDatastore_Task, 'duration_secs': 0.011264} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.148989] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.149320] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 2259c42a-79f8-455f-b29f-464705526932/2259c42a-79f8-455f-b29f-464705526932.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.149644] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46788923-4f29-4c82-98d8-d461deed528b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.158676] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 919.158676] env[61006]: value = "task-1337246" [ 919.158676] env[61006]: _type = "Task" [ 919.158676] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.161969] env[61006]: DEBUG nova.virt.block_device [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating existing volume attachment record: d8e98ceb-bab3-4148-b070-afc5b81937f0 {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 919.172653] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.207477] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.396893] env[61006]: DEBUG nova.network.neutron [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.415299] env[61006]: DEBUG nova.scheduler.client.report [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.489869] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337245, 'name': ReconfigVM_Task, 'duration_secs': 0.463622} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.490284] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 1d6ede6a-5601-4da4-baa1-5090a112ec2d/1d6ede6a-5601-4da4-baa1-5090a112ec2d.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.490973] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d6f660f-5492-4930-9fd0-6f8d250a84dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.500651] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 919.500651] env[61006]: value = "task-1337248" [ 919.500651] env[61006]: _type = "Task" [ 919.500651] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.511023] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337248, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.555070] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337244, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.597986] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.625630] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.625965] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.626172] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.626367] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.626540] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.626700] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.627010] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.627242] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.627440] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.627616] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.627817] env[61006]: DEBUG nova.virt.hardware [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.628786] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f302f878-cd20-4944-86a4-5340187b383e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.638604] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cd6eb4-39e1-4aa7-a33a-76bdcef46aee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.673029] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337246, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.900512] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.900965] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Instance network_info: |[{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.901326] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.901530] env[61006]: DEBUG nova.network.neutron [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.902958] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:9e:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58bb0db5-4669-4185-8d20-b5a77724df9b', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.910990] env[61006]: DEBUG oslo.service.loopingcall [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.912877] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.913298] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fe54b20-330b-43a0-9210-846bafbb4189 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.935024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.935024] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 919.935844] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.744s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.936092] env[61006]: DEBUG nova.objects.instance [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lazy-loading 'resources' on Instance uuid e2a40bd9-fb66-40a2-bcf1-5c74707d59dd {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.945333] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.945333] env[61006]: value = "task-1337249" [ 919.945333] env[61006]: _type = "Task" [ 919.945333] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.959132] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337249, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.011340] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337248, 'name': Rename_Task, 'duration_secs': 0.311548} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.011636] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.012097] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4458440-f5e4-45ab-972d-73ec1d3d0763 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.020109] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 920.020109] env[61006]: value = "task-1337250" [ 920.020109] env[61006]: _type = "Task" [ 920.020109] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.030632] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337250, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.056073] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337244, 'name': CloneVM_Task} progress is 95%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.081416] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Successfully updated port: 7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.175630] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5857} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.176663] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 2259c42a-79f8-455f-b29f-464705526932/2259c42a-79f8-455f-b29f-464705526932.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.176663] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.176663] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d4999af-d17a-4a64-98d4-c3803ce7b530 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.184538] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 920.184538] env[61006]: value = "task-1337251" [ 920.184538] env[61006]: _type = "Task" [ 920.184538] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.196847] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337251, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.443844] env[61006]: DEBUG nova.compute.utils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.447663] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.447861] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.462790] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337249, 'name': CreateVM_Task, 'duration_secs': 0.418915} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.462970] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.463666] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.463888] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.464155] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.467142] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65351b4e-d37b-4845-bed5-dccef19ac75b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.475136] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 920.475136] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b25e7-5438-2728-3505-9294aefcbae4" [ 920.475136] env[61006]: _type = "Task" [ 920.475136] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.485145] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b25e7-5438-2728-3505-9294aefcbae4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.538139] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337250, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.543764] env[61006]: DEBUG nova.policy [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.565891] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337244, 'name': CloneVM_Task, 'duration_secs': 1.667311} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.570287] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Created linked-clone VM from snapshot [ 920.571980] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7942bff-8224-4ae1-9a3a-78d710481300 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.581443] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Uploading image 87693d49-a593-43f4-8a3e-5dc9080f7010 {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 920.585324] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.585374] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.585662] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.616246] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 920.616246] env[61006]: value = "vm-285383" [ 920.616246] env[61006]: _type = "VirtualMachine" [ 920.616246] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 920.616819] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b67d50b1-7f21-47fb-b429-6cd3d58404a1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.628400] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease: (returnval){ [ 920.628400] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be8a09-e111-e90a-cb0b-4b037937a8c0" [ 920.628400] env[61006]: _type = "HttpNfcLease" [ 920.628400] env[61006]: } obtained for exporting VM: (result){ [ 920.628400] env[61006]: value = "vm-285383" [ 920.628400] env[61006]: _type = "VirtualMachine" [ 920.628400] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 920.628752] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the lease: (returnval){ [ 920.628752] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be8a09-e111-e90a-cb0b-4b037937a8c0" [ 920.628752] env[61006]: _type = "HttpNfcLease" [ 920.628752] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 920.640215] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 920.640215] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be8a09-e111-e90a-cb0b-4b037937a8c0" [ 920.640215] env[61006]: _type = "HttpNfcLease" [ 920.640215] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 920.696082] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337251, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081164} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.698577] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.699824] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66595aab-8627-46ec-b670-83d63d79031c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.724052] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 2259c42a-79f8-455f-b29f-464705526932/2259c42a-79f8-455f-b29f-464705526932.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.729161] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48c68158-b50d-4279-865d-7fd8323ff21f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.747570] env[61006]: DEBUG nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.747570] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c002e0-cd4f-4cc7-a495-0d5328aa1ad8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.758987] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 920.758987] env[61006]: value = "task-1337253" [ 920.758987] env[61006]: _type = "Task" [ 920.758987] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.771515] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.775837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c7c57c-ca0f-4280-b2ed-462a97635c0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.784281] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6663ccb1-980d-42ad-9919-c5d61c151545 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.816369] env[61006]: DEBUG nova.network.neutron [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.816730] env[61006]: DEBUG nova.network.neutron [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.818534] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a01ec73-ddd7-445c-8ce9-8b7151cf9ad3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.829367] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b620343c-41a7-4425-82df-a4c261df4e06 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.844951] env[61006]: DEBUG nova.compute.provider_tree [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.935652] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Successfully created port: 23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.950833] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 920.987197] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b25e7-5438-2728-3505-9294aefcbae4, 'name': SearchDatastore_Task, 'duration_secs': 0.012262} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.987666] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.987778] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.988022] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.988190] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.988371] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.988662] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d4830d9-b90a-470c-bc7c-70959c2a2971 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.998627] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.998847] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.999916] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39a1af6c-dd37-4520-bbf1-a2d5cbfe606e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.006614] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 921.006614] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd60cd-23a9-358e-25ee-b18f42dc695a" [ 921.006614] env[61006]: _type = "Task" [ 921.006614] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.017455] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd60cd-23a9-358e-25ee-b18f42dc695a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.027014] env[61006]: DEBUG nova.compute.manager [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Received event network-vif-plugged-7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.027243] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Acquiring lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.027464] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.027601] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.027766] env[61006]: DEBUG nova.compute.manager [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] No waiting events found dispatching network-vif-plugged-7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.027932] env[61006]: WARNING nova.compute.manager [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Received unexpected event network-vif-plugged-7fe1873c-f1dd-498a-893b-b179d4714d5d for instance with vm_state building and task_state spawning. [ 921.028151] env[61006]: DEBUG nova.compute.manager [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Received event network-changed-7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.028244] env[61006]: DEBUG nova.compute.manager [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Refreshing instance network info cache due to event network-changed-7fe1873c-f1dd-498a-893b-b179d4714d5d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.028691] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Acquiring lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.035259] env[61006]: DEBUG oslo_vmware.api [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337250, 'name': PowerOnVM_Task, 'duration_secs': 0.734107} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.035548] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.035881] env[61006]: INFO nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Took 8.64 seconds to spawn the instance on the hypervisor. [ 921.035987] env[61006]: DEBUG nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.037056] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f6fdbc-fe80-4193-add4-5ca7b66d821b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.134212] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.140305] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.140305] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be8a09-e111-e90a-cb0b-4b037937a8c0" [ 921.140305] env[61006]: _type = "HttpNfcLease" [ 921.140305] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 921.140738] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 921.140738] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52be8a09-e111-e90a-cb0b-4b037937a8c0" [ 921.140738] env[61006]: _type = "HttpNfcLease" [ 921.140738] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 921.141499] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894a7a78-98dc-4b8d-b9aa-7220a361b296 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.150066] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 921.150271] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 921.260727] env[61006]: INFO nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] instance snapshotting [ 921.261439] env[61006]: DEBUG nova.objects.instance [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.263207] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-caaea245-f256-4cc5-b5ab-d26d7b35aa05 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.274578] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337253, 'name': ReconfigVM_Task, 'duration_secs': 0.347729} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.276185] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 2259c42a-79f8-455f-b29f-464705526932/2259c42a-79f8-455f-b29f-464705526932.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.278144] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07941a21-e79a-4296-bd2d-5866194c45da {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.292416] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 921.292416] env[61006]: value = "task-1337254" [ 921.292416] env[61006]: _type = "Task" [ 921.292416] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.311194] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337254, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.322587] env[61006]: DEBUG oslo_concurrency.lockutils [req-b976e47a-1b4f-4729-995a-80c6e34bb5d4 req-60b90752-b46e-485d-aff1-df15fe759125 service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.348483] env[61006]: DEBUG nova.scheduler.client.report [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.455058] env[61006]: DEBUG nova.network.neutron [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Updating instance_info_cache with network_info: [{"id": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "address": "fa:16:3e:e5:6d:19", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe1873c-f1", "ovs_interfaceid": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.525519] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52dd60cd-23a9-358e-25ee-b18f42dc695a, 'name': SearchDatastore_Task, 'duration_secs': 0.012073} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.526369] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16972d38-8a45-46ef-bbbb-af102f5b86d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.533288] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 921.533288] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527fd7a8-3d0c-1133-94bf-d2ff8981027b" [ 921.533288] env[61006]: _type = "Task" [ 921.533288] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.545167] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527fd7a8-3d0c-1133-94bf-d2ff8981027b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.559466] env[61006]: INFO nova.compute.manager [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Took 20.94 seconds to build instance. [ 921.769629] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f7f21d-3974-45b4-bba7-a9a84baf1956 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.798837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3586a52e-fd1e-488d-a2ec-8e3e08e12ac8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.808786] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337254, 'name': Rename_Task, 'duration_secs': 0.221867} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.811048] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.813928] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40ffffb1-224f-4011-8448-f5edba49952c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.824648] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 921.824648] env[61006]: value = "task-1337256" [ 921.824648] env[61006]: _type = "Task" [ 921.824648] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.835765] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337256, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.853973] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.857109] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.099s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.857446] env[61006]: DEBUG nova.objects.instance [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lazy-loading 'resources' on Instance uuid c5d6f6df-d296-4d4d-815c-159dad2d08dc {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.882680] env[61006]: INFO nova.scheduler.client.report [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted allocations for instance e2a40bd9-fb66-40a2-bcf1-5c74707d59dd [ 921.962457] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.962842] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Instance network_info: |[{"id": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "address": "fa:16:3e:e5:6d:19", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe1873c-f1", "ovs_interfaceid": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 921.964109] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Acquired lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.964527] env[61006]: DEBUG nova.network.neutron [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Refreshing network info cache for port 7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.966851] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 921.970879] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e5:6d:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1931669-8959-4e86-a603-e206bcf2b47a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fe1873c-f1dd-498a-893b-b179d4714d5d', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.982069] env[61006]: DEBUG oslo.service.loopingcall [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.984060] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.984530] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37df1fba-c8e4-4d80-9b74-d6b6e30900f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.007748] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.007748] env[61006]: value = "task-1337257" [ 922.007748] env[61006]: _type = "Task" [ 922.007748] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.016849] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337257, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.045266] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527fd7a8-3d0c-1133-94bf-d2ff8981027b, 'name': SearchDatastore_Task, 'duration_secs': 0.01297} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.045266] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.045381] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a/bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.045733] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0897566e-ed2c-45da-bc57-486b1d9ffa92 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.053398] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 922.053398] env[61006]: value = "task-1337258" [ 922.053398] env[61006]: _type = "Task" [ 922.053398] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.063601] env[61006]: DEBUG oslo_concurrency.lockutils [None req-dda0238c-f4cf-4d79-94b5-1a3cd76728a0 tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.454s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.064255] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337258, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.319377] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 922.319716] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-894559fc-e794-4335-952c-d7ddf18760f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.332403] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 922.332403] env[61006]: value = "task-1337259" [ 922.332403] env[61006]: _type = "Task" [ 922.332403] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.341656] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337256, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.345740] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337259, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.394718] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6a16417-b13e-406f-afed-42d3ba8a6740 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "e2a40bd9-fb66-40a2-bcf1-5c74707d59dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.708s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.520616] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337257, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.570240] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337258, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.745544] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8642ed1b-223e-4585-ac48-1920c6e28b0a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.760405] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd593a5-de3f-4b1d-8282-be436b7129d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.797332] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01f2bb8-391b-47da-ae5c-20790f463462 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.809524] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754b4e58-b493-4ecf-a757-dc5c1bce07aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.827503] env[61006]: DEBUG nova.compute.provider_tree [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.841616] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337256, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.847995] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337259, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.965663] env[61006]: DEBUG nova.network.neutron [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Updated VIF entry in instance network info cache for port 7fe1873c-f1dd-498a-893b-b179d4714d5d. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.966161] env[61006]: DEBUG nova.network.neutron [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Updating instance_info_cache with network_info: [{"id": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "address": "fa:16:3e:e5:6d:19", "network": {"id": "677aa294-6dd3-4673-9d9b-659d364fc242", "bridge": "br-int", "label": "tempest-ImagesTestJSON-426009608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d26132a67fc1412c938a36094a34a433", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1931669-8959-4e86-a603-e206bcf2b47a", "external-id": "nsx-vlan-transportzone-937", "segmentation_id": 937, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fe1873c-f1", "ovs_interfaceid": "7fe1873c-f1dd-498a-893b-b179d4714d5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.021435] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337257, 'name': CreateVM_Task, 'duration_secs': 0.643753} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.021635] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.022453] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.022710] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.023351] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.023621] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-561c646a-e449-40ac-9282-686985477a94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.030179] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 923.030179] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52861d2e-398a-6997-61ad-175acb2d0fa0" [ 923.030179] env[61006]: _type = "Task" [ 923.030179] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.039440] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52861d2e-398a-6997-61ad-175acb2d0fa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.067077] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337258, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670122} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.067465] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a/bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.067757] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.068102] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-987f0500-539e-4f86-858b-fa76fdea44f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.076701] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 923.076701] env[61006]: value = "task-1337260" [ 923.076701] env[61006]: _type = "Task" [ 923.076701] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.087562] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.277643] env[61006]: DEBUG nova.compute.manager [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Received event network-vif-plugged-23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.277955] env[61006]: DEBUG oslo_concurrency.lockutils [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] Acquiring lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.281026] env[61006]: DEBUG oslo_concurrency.lockutils [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.281244] env[61006]: DEBUG oslo_concurrency.lockutils [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.281533] env[61006]: DEBUG nova.compute.manager [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] No waiting events found dispatching network-vif-plugged-23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.281743] env[61006]: WARNING nova.compute.manager [req-1b511ba0-95a1-4e49-a20e-4dd65881d7b7 req-57098a73-10ba-4792-bb05-926af8fefcff service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Received unexpected event network-vif-plugged-23bdf050-f167-4c83-a7c5-8b64df012cd0 for instance with vm_state building and task_state spawning. [ 923.334262] env[61006]: DEBUG nova.scheduler.client.report [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.342563] env[61006]: DEBUG oslo_vmware.api [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337256, 'name': PowerOnVM_Task, 'duration_secs': 1.030159} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.346567] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.347199] env[61006]: INFO nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Took 8.60 seconds to spawn the instance on the hypervisor. [ 923.347504] env[61006]: DEBUG nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.348881] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d77ffb0-0aa3-4d73-9898-bf86c6069e99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.357394] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337259, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.453422] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Successfully updated port: 23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.468651] env[61006]: DEBUG oslo_concurrency.lockutils [req-20e2e3e4-7f07-458f-b8b8-d29569ed3717 req-2985d9bb-a210-4606-bf35-3af3bd6be902 service nova] Releasing lock "refresh_cache-40500a0d-33f3-492a-9573-ed10e45642c5" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.544560] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52861d2e-398a-6997-61ad-175acb2d0fa0, 'name': SearchDatastore_Task, 'duration_secs': 0.016281} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.545098] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.545490] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.545839] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.546026] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.546274] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.546673] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f710acd-cc90-4ac2-9543-0dc3c5617596 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.566632] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.566881] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.567706] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4dcf8a7-d6fa-4b58-aa8b-c197dc9c084f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.574516] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 923.574516] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520688fe-abe8-738c-cd6c-2d0576a535f9" [ 923.574516] env[61006]: _type = "Task" [ 923.574516] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.588884] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264992} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.592103] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.592532] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520688fe-abe8-738c-cd6c-2d0576a535f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.593402] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4085ca-ce9d-4d9a-9561-ab3b7a0b2db4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.619397] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a/bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.621034] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4921edd0-8eeb-4b51-a29a-a42d94d240e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.637109] env[61006]: DEBUG nova.compute.manager [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Received event network-changed-23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.637321] env[61006]: DEBUG nova.compute.manager [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Refreshing instance network info cache due to event network-changed-23bdf050-f167-4c83-a7c5-8b64df012cd0. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.637576] env[61006]: DEBUG oslo_concurrency.lockutils [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] Acquiring lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.637724] env[61006]: DEBUG oslo_concurrency.lockutils [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] Acquired lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.637905] env[61006]: DEBUG nova.network.neutron [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Refreshing network info cache for port 23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.646877] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 923.646877] env[61006]: value = "task-1337261" [ 923.646877] env[61006]: _type = "Task" [ 923.646877] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.660472] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337261, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.712866] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 923.713125] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285384', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'name': 'volume-9f7e0232-99d9-446e-aca8-77c70e411002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'serial': '9f7e0232-99d9-446e-aca8-77c70e411002'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 923.714109] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc165d10-c8dd-4ccb-bb92-c6ed53ea60e4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.735322] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1f3c32-f3ad-43ab-8a68-0969e5579ab6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.766878] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] volume-9f7e0232-99d9-446e-aca8-77c70e411002/volume-9f7e0232-99d9-446e-aca8-77c70e411002.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.767322] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e5f7073-effc-41a3-9dbf-622e2d32f7a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.789341] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 923.789341] env[61006]: value = "task-1337262" [ 923.789341] env[61006]: _type = "Task" [ 923.789341] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.799369] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.847657] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.850220] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 11.041s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.858237] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337259, 'name': CreateSnapshot_Task, 'duration_secs': 1.274562} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.858505] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 923.859412] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63149f5c-e1e9-4bcf-808d-953e20117ff2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.877977] env[61006]: INFO nova.compute.manager [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Took 22.82 seconds to build instance. [ 923.882755] env[61006]: INFO nova.scheduler.client.report [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Deleted allocations for instance c5d6f6df-d296-4d4d-815c-159dad2d08dc [ 923.957074] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.086686] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520688fe-abe8-738c-cd6c-2d0576a535f9, 'name': SearchDatastore_Task, 'duration_secs': 0.018147} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.087598] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb438f26-35b6-4aa4-985b-93f237d40225 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.095214] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 924.095214] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52102c33-9ac8-5bf7-368e-988fac24fc9e" [ 924.095214] env[61006]: _type = "Task" [ 924.095214] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.105287] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52102c33-9ac8-5bf7-368e-988fac24fc9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.157689] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337261, 'name': ReconfigVM_Task, 'duration_secs': 0.489007} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.157983] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfigured VM instance instance-00000059 to attach disk [datastore2] bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a/bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.158685] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90392718-2e0d-4d8a-ba91-e48c0de668eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.166989] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 924.166989] env[61006]: value = "task-1337263" [ 924.166989] env[61006]: _type = "Task" [ 924.166989] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.178069] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337263, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.197214] env[61006]: DEBUG nova.network.neutron [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 924.301106] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.316711] env[61006]: DEBUG nova.network.neutron [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.385638] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 924.386517] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4756f72d-2376-48b6-b445-e8cece5d6536 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.337s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.386771] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f85c3bcd-ca0b-4eeb-a879-69711419d08b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.393557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-112ca612-bf2d-47df-891d-9a7e655c14ad tempest-InstanceActionsTestJSON-1639360245 tempest-InstanceActionsTestJSON-1639360245-project-member] Lock "c5d6f6df-d296-4d4d-815c-159dad2d08dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.535s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.402734] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 924.402734] env[61006]: value = "task-1337264" [ 924.402734] env[61006]: _type = "Task" [ 924.402734] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.415083] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337264, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.501518] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.501801] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.502236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.502469] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.503038] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.508742] env[61006]: INFO nova.compute.manager [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Terminating instance [ 924.513090] env[61006]: DEBUG nova.compute.manager [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.513090] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.513294] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b840cd49-b676-40b7-bc8b-d8543f118842 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.526853] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.527253] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d00d0eb-dedc-4801-a4b4-cbeb8f5c292d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.535356] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 924.535356] env[61006]: value = "task-1337265" [ 924.535356] env[61006]: _type = "Task" [ 924.535356] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.550836] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.603612] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1da109-b03e-4079-bd86-210bc9ff514f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.614535] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52102c33-9ac8-5bf7-368e-988fac24fc9e, 'name': SearchDatastore_Task, 'duration_secs': 0.018868} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.616972] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.617329] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 40500a0d-33f3-492a-9573-ed10e45642c5/40500a0d-33f3-492a-9573-ed10e45642c5.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.617676] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c181b3a9-93d7-428f-882a-02123e288213 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.620750] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3e74d2-238f-43f8-8208-984c2ff35a09 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.658050] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a58cbe-2610-4f13-8fa9-cd78c63e94c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.662224] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 924.662224] env[61006]: value = "task-1337266" [ 924.662224] env[61006]: _type = "Task" [ 924.662224] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.675462] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50af04e0-cfeb-4c34-bd31-7181741300bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.683351] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.689586] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337263, 'name': Rename_Task, 'duration_secs': 0.223845} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.699794] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.700532] env[61006]: DEBUG nova.compute.provider_tree [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.702200] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17619d2b-2e36-424c-95f5-597445181b15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.713505] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 924.713505] env[61006]: value = "task-1337267" [ 924.713505] env[61006]: _type = "Task" [ 924.713505] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.725468] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337267, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.754148] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.754453] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.754663] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.754972] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.755095] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.755322] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.755626] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.755945] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.756118] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.756296] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.757047] env[61006]: DEBUG nova.virt.hardware [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.759664] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c23960e-b18a-4e46-8f21-72961a66453f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.771307] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af67257d-c9c2-448c-a2d0-b29765610b75 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.778086] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 924.779070] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1864d4d-1ebf-45b7-b2a0-a85f13c3e2de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.799598] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 924.799838] env[61006]: ERROR oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk due to incomplete transfer. [ 924.800201] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c63e425a-0e37-4d04-b777-e2a078d6ae32 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.810167] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337262, 'name': ReconfigVM_Task, 'duration_secs': 0.673528} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.810480] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfigured VM instance instance-0000004c to attach disk [datastore2] volume-9f7e0232-99d9-446e-aca8-77c70e411002/volume-9f7e0232-99d9-446e-aca8-77c70e411002.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.815890] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d0994d6-e078-48bc-9b2e-028001af2101 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.828940] env[61006]: DEBUG oslo_concurrency.lockutils [req-56481dd6-02a3-4848-ba0c-50419c664f7c req-146d987e-74fb-4647-9e1e-6f521f0f835e service nova] Releasing lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.829420] env[61006]: DEBUG oslo_vmware.rw_handles [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f97f64-0c1a-6d1a-0796-ab1f2ef92c54/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 924.829622] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Uploaded image 07e07d31-cef3-4ce6-a79e-0a94d33e19b9 to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 924.832050] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 924.832197] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.832313] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.833994] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-26e3b66b-564f-4164-86f9-d0c0bfed221f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.851210] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 924.851210] env[61006]: value = "task-1337268" [ 924.851210] env[61006]: _type = "Task" [ 924.851210] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.851702] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 924.851702] env[61006]: value = "task-1337269" [ 924.851702] env[61006]: _type = "Task" [ 924.851702] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.871437] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337268, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.876447] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337269, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.914872] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337264, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.047978] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337265, 'name': PowerOffVM_Task, 'duration_secs': 0.451418} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.048344] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.048551] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.048887] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14e53735-163b-400a-88ea-dc475cb9e520 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.144204] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.144723] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.145016] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Deleting the datastore file [datastore2] 1d6ede6a-5601-4da4-baa1-5090a112ec2d {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.145466] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4448da2-8f3a-4986-9874-a890af0e67dc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.154964] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for the task: (returnval){ [ 925.154964] env[61006]: value = "task-1337271" [ 925.154964] env[61006]: _type = "Task" [ 925.154964] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.171342] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.181531] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337266, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.206024] env[61006]: DEBUG nova.scheduler.client.report [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.225269] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337267, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.365129] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318fd302-8388-473d-8a7b-984280792338 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.374218] env[61006]: DEBUG oslo_vmware.api [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337268, 'name': ReconfigVM_Task, 'duration_secs': 0.185512} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.380083] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.382452] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285384', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'name': 'volume-9f7e0232-99d9-446e-aca8-77c70e411002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'serial': '9f7e0232-99d9-446e-aca8-77c70e411002'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 925.384327] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337269, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.385387] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Suspending the VM {{(pid=61006) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 925.385387] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-37eaaeac-417d-4218-96fb-204df9fc0e89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.394600] env[61006]: DEBUG oslo_vmware.api [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 925.394600] env[61006]: value = "task-1337272" [ 925.394600] env[61006]: _type = "Task" [ 925.394600] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.411131] env[61006]: DEBUG oslo_vmware.api [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337272, 'name': SuspendVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.420213] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337264, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.656852] env[61006]: DEBUG nova.network.neutron [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Updating instance_info_cache with network_info: [{"id": "23bdf050-f167-4c83-a7c5-8b64df012cd0", "address": "fa:16:3e:0a:14:cf", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23bdf050-f1", "ovs_interfaceid": "23bdf050-f167-4c83-a7c5-8b64df012cd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.675071] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.679544] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337266, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.814303} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.679913] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 40500a0d-33f3-492a-9573-ed10e45642c5/40500a0d-33f3-492a-9573-ed10e45642c5.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.680113] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.680660] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-26c1aa07-cee7-4067-8344-263895e191c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.689902] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 925.689902] env[61006]: value = "task-1337273" [ 925.689902] env[61006]: _type = "Task" [ 925.689902] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.701534] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337273, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.730669] env[61006]: DEBUG oslo_vmware.api [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337267, 'name': PowerOnVM_Task, 'duration_secs': 0.829633} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.732176] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.732326] env[61006]: INFO nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Took 8.56 seconds to spawn the instance on the hypervisor. [ 925.732424] env[61006]: DEBUG nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.734161] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5355f176-a58f-4f06-b611-bba71a8625c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.866453] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337269, 'name': Destroy_Task, 'duration_secs': 0.792089} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.867872] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Destroyed the VM [ 925.868812] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 925.868812] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1c9b6463-28fb-4d0c-85b4-1199b4778387 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.878590] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 925.878590] env[61006]: value = "task-1337274" [ 925.878590] env[61006]: _type = "Task" [ 925.878590] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.894271] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337274, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.907473] env[61006]: DEBUG oslo_vmware.api [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337272, 'name': SuspendVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.918966] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337264, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.165290] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.165825] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance network_info: |[{"id": "23bdf050-f167-4c83-a7c5-8b64df012cd0", "address": "fa:16:3e:0a:14:cf", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23bdf050-f1", "ovs_interfaceid": "23bdf050-f167-4c83-a7c5-8b64df012cd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.172016] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:14:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23bdf050-f167-4c83-a7c5-8b64df012cd0', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.179827] env[61006]: DEBUG oslo.service.loopingcall [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.180432] env[61006]: DEBUG oslo_vmware.api [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Task: {'id': task-1337271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.69939} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.180906] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.181313] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.181630] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.182027] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.182356] env[61006]: INFO nova.compute.manager [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Took 1.67 seconds to destroy the instance on the hypervisor. [ 926.184017] env[61006]: DEBUG oslo.service.loopingcall [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.184017] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ca50ce8-9d1f-4233-b71b-4f7c94d87b57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.206167] env[61006]: DEBUG nova.compute.manager [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.206346] env[61006]: DEBUG nova.network.neutron [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 926.220556] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337273, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109316} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.221773] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.222152] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.222152] env[61006]: value = "task-1337275" [ 926.222152] env[61006]: _type = "Task" [ 926.222152] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.223478] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.373s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.230023] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027a7473-72a4-4adc-a518-efccf09c4f35 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.241373] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337275, 'name': CreateVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.267381] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 40500a0d-33f3-492a-9573-ed10e45642c5/40500a0d-33f3-492a-9573-ed10e45642c5.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.269839] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8231d85-ac1c-47a4-bf24-27fb186b0317 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.289433] env[61006]: INFO nova.compute.manager [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Took 22.09 seconds to build instance. [ 926.296967] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 926.296967] env[61006]: value = "task-1337276" [ 926.296967] env[61006]: _type = "Task" [ 926.296967] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.307367] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.389298] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337274, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.408271] env[61006]: DEBUG oslo_vmware.api [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337272, 'name': SuspendVM_Task, 'duration_secs': 0.727031} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.412327] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Suspended the VM {{(pid=61006) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 926.412641] env[61006]: DEBUG nova.compute.manager [None req-095dadc9-03b6-497d-90bd-2b3a9e7da95d tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.413546] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec29e7d-56a5-41bf-804f-ed0a8c2b2cca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.423113] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337264, 'name': CloneVM_Task, 'duration_secs': 1.740084} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.425215] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Created linked-clone VM from snapshot [ 926.433022] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67abb6a5-fa6b-459c-83a1-5a7a1ac9d266 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.440251] env[61006]: DEBUG nova.objects.instance [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lazy-loading 'flavor' on Instance uuid d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.442389] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploading image 767836f1-ca1e-4962-9506-b705f9ab4d4c {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 926.471085] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 926.471085] env[61006]: value = "vm-285388" [ 926.471085] env[61006]: _type = "VirtualMachine" [ 926.471085] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 926.472444] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-903df380-9147-4886-b88d-966f223d0bbf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.485576] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease: (returnval){ [ 926.485576] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058c5-2661-bcd1-8449-02dbe1c94466" [ 926.485576] env[61006]: _type = "HttpNfcLease" [ 926.485576] env[61006]: } obtained for exporting VM: (result){ [ 926.485576] env[61006]: value = "vm-285388" [ 926.485576] env[61006]: _type = "VirtualMachine" [ 926.485576] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 926.485576] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the lease: (returnval){ [ 926.485576] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058c5-2661-bcd1-8449-02dbe1c94466" [ 926.485576] env[61006]: _type = "HttpNfcLease" [ 926.485576] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 926.493855] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.493855] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058c5-2661-bcd1-8449-02dbe1c94466" [ 926.493855] env[61006]: _type = "HttpNfcLease" [ 926.493855] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 926.750976] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337275, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.792370] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a0e385ea-2bd5-4d03-a073-cee551f1e749 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.603s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.810124] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337276, 'name': ReconfigVM_Task, 'duration_secs': 0.36031} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.810521] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 40500a0d-33f3-492a-9573-ed10e45642c5/40500a0d-33f3-492a-9573-ed10e45642c5.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.811314] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d3d6787-7156-4db2-8838-10c75f352b33 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.819167] env[61006]: INFO nova.scheduler.client.report [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocation for migration 6d25c2c7-a5cf-4f93-b8eb-0de8b6265269 [ 926.824377] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 926.824377] env[61006]: value = "task-1337278" [ 926.824377] env[61006]: _type = "Task" [ 926.824377] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.826387] env[61006]: DEBUG nova.compute.manager [req-b3dda693-a4ac-481b-a892-a6bfada48c9f req-77da0785-edaf-4646-92e4-cd5a6750a620 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Received event network-vif-deleted-beb7410d-8ed5-460f-937c-81ace8d6600c {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.827346] env[61006]: INFO nova.compute.manager [req-b3dda693-a4ac-481b-a892-a6bfada48c9f req-77da0785-edaf-4646-92e4-cd5a6750a620 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Neutron deleted interface beb7410d-8ed5-460f-937c-81ace8d6600c; detaching it from the instance and deleting it from the info cache [ 926.827731] env[61006]: DEBUG nova.network.neutron [req-b3dda693-a4ac-481b-a892-a6bfada48c9f req-77da0785-edaf-4646-92e4-cd5a6750a620 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.843744] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337278, 'name': Rename_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.895030] env[61006]: DEBUG oslo_vmware.api [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337274, 'name': RemoveSnapshot_Task, 'duration_secs': 0.800391} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.895030] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 926.895030] env[61006]: INFO nova.compute.manager [None req-044e66d5-dfea-41c1-9843-dd867ff21f9f tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 16.41 seconds to snapshot the instance on the hypervisor. [ 926.947408] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f429047e-0a0f-43f9-921b-568a58148162 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.853s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.994027] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.994027] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058c5-2661-bcd1-8449-02dbe1c94466" [ 926.994027] env[61006]: _type = "HttpNfcLease" [ 926.994027] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 926.994027] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 926.994027] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]520058c5-2661-bcd1-8449-02dbe1c94466" [ 926.994027] env[61006]: _type = "HttpNfcLease" [ 926.994027] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 926.994745] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611381c0-5996-4c82-8c90-5dd202fd1936 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.003097] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 927.003097] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 927.106097] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d39b78f3-0250-4173-9f3c-2d6e187dae86 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.108140] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.108391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.231542] env[61006]: DEBUG nova.compute.manager [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.231542] env[61006]: DEBUG nova.compute.manager [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.231542] env[61006]: DEBUG oslo_concurrency.lockutils [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.231542] env[61006]: DEBUG oslo_concurrency.lockutils [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.231542] env[61006]: DEBUG nova.network.neutron [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.249982] env[61006]: DEBUG nova.network.neutron [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.252156] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337275, 'name': CreateVM_Task, 'duration_secs': 0.541561} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.252156] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.252679] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.252835] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.253180] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.253660] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14196fae-a63b-4503-b071-ab74317b475b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.261387] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 927.261387] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528000be-3db8-f1f8-e558-a7d4d4c37451" [ 927.261387] env[61006]: _type = "Task" [ 927.261387] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.273686] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528000be-3db8-f1f8-e558-a7d4d4c37451, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.339113] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e43ad4c0-8a1f-493c-8444-4b364b44e28f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.344182] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f9d39439-8864-4428-b538-7481bc1fbfc2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 18.104s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.351771] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337278, 'name': Rename_Task, 'duration_secs': 0.196909} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.353336] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.353674] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-add2c03a-41f4-4732-9266-36619b51d0be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.358383] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332c9f4d-d989-44ca-8e94-652ca68aab57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.379283] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 927.379283] env[61006]: value = "task-1337279" [ 927.379283] env[61006]: _type = "Task" [ 927.379283] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.397277] env[61006]: DEBUG nova.compute.manager [req-b3dda693-a4ac-481b-a892-a6bfada48c9f req-77da0785-edaf-4646-92e4-cd5a6750a620 service nova] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Detach interface failed, port_id=beb7410d-8ed5-460f-937c-81ace8d6600c, reason: Instance 1d6ede6a-5601-4da4-baa1-5090a112ec2d could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 927.403053] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.612206] env[61006]: INFO nova.compute.manager [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Detaching volume e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83 [ 927.667474] env[61006]: INFO nova.virt.block_device [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attempting to driver detach volume e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83 from mountpoint /dev/sdb [ 927.668358] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 927.669239] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285375', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'name': 'volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'serial': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 927.670474] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e598dc10-a051-4f08-9634-2d3c37d146e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.709393] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82c7d1c-84cb-47bd-b2b9-002edbbfe411 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.724292] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30088583-29bd-42e7-b8a2-bd7f9599b564 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.760606] env[61006]: INFO nova.compute.manager [-] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Took 1.55 seconds to deallocate network for instance. [ 927.765806] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12accc1-4478-467e-957b-cadd934d4911 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.793170] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528000be-3db8-f1f8-e558-a7d4d4c37451, 'name': SearchDatastore_Task, 'duration_secs': 0.012597} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.806979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.807088] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.807344] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.808135] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.808135] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.809034] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] The volume has not been displaced from its original location: [datastore2] volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83/volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 927.819067] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 927.822836] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8a508e9-e902-46ec-b133-140e02188637 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.822836] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8452c8c-8f82-46c3-ab82-6a6de50cdf7c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.847252] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 927.847252] env[61006]: value = "task-1337280" [ 927.847252] env[61006]: _type = "Task" [ 927.847252] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.851883] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.852289] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.853701] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b5d685f-4a15-4e46-8d1c-9cf25123af1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.860486] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.865830] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 927.865830] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52715cf4-0f3d-c56b-1078-fd0219b0be81" [ 927.865830] env[61006]: _type = "Task" [ 927.865830] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.879325] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52715cf4-0f3d-c56b-1078-fd0219b0be81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.893954] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337279, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.280593] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "2259c42a-79f8-455f-b29f-464705526932" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.280952] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.281635] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "2259c42a-79f8-455f-b29f-464705526932-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.281877] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.282496] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.286529] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.286775] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.286996] env[61006]: DEBUG nova.objects.instance [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lazy-loading 'resources' on Instance uuid 1d6ede6a-5601-4da4-baa1-5090a112ec2d {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.288314] env[61006]: INFO nova.compute.manager [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Terminating instance [ 928.292254] env[61006]: DEBUG nova.compute.manager [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.292588] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.293504] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5bdaa6-94e1-4a23-b290-dd71d67a830c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.309222] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.309609] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.314538] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.315805] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a970ed3-234d-472e-aefa-1ed48c6e7413 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.361539] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337280, 'name': ReconfigVM_Task, 'duration_secs': 0.375758} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.365702] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.371782] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5212178a-c5fc-491c-b308-ff12725a40cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.402596] env[61006]: DEBUG oslo_vmware.api [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337279, 'name': PowerOnVM_Task, 'duration_secs': 0.545263} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.408108] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.408614] env[61006]: INFO nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Took 8.81 seconds to spawn the instance on the hypervisor. [ 928.408861] env[61006]: DEBUG nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.409497] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 928.409497] env[61006]: value = "task-1337282" [ 928.409497] env[61006]: _type = "Task" [ 928.409497] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.409892] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.410235] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.410467] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleting the datastore file [datastore2] 2259c42a-79f8-455f-b29f-464705526932 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.410772] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52715cf4-0f3d-c56b-1078-fd0219b0be81, 'name': SearchDatastore_Task, 'duration_secs': 0.013594} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.411645] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadd0256-907f-4d6e-b96c-a7686f0d1dec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.414863] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fba50228-7a9c-4f15-8188-b36e4192e469 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.421218] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18e767c8-7fe5-4e8c-8990-4495305b333f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.440527] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337282, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.441068] env[61006]: DEBUG oslo_vmware.api [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 928.441068] env[61006]: value = "task-1337283" [ 928.441068] env[61006]: _type = "Task" [ 928.441068] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.441601] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 928.441601] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c5651f-95a8-d408-fba7-8d749ffb5688" [ 928.441601] env[61006]: _type = "Task" [ 928.441601] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.461921] env[61006]: DEBUG oslo_vmware.api [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.466851] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c5651f-95a8-d408-fba7-8d749ffb5688, 'name': SearchDatastore_Task, 'duration_secs': 0.029344} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.467261] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.467630] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1/97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.467949] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7d555f0-8f85-40ce-b68b-635a6a8e4210 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.479531] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 928.479531] env[61006]: value = "task-1337284" [ 928.479531] env[61006]: _type = "Task" [ 928.479531] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.490376] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337284, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.700209] env[61006]: DEBUG nova.network.neutron [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.700608] env[61006]: DEBUG nova.network.neutron [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.816574] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 928.933408] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337282, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.974298] env[61006]: INFO nova.compute.manager [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Took 23.80 seconds to build instance. [ 928.980792] env[61006]: DEBUG oslo_vmware.api [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316161} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.985695] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.985695] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.986072] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.986334] env[61006]: INFO nova.compute.manager [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: 2259c42a-79f8-455f-b29f-464705526932] Took 0.69 seconds to destroy the instance on the hypervisor. [ 928.986697] env[61006]: DEBUG oslo.service.loopingcall [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.987073] env[61006]: DEBUG nova.compute.manager [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.987408] env[61006]: DEBUG nova.network.neutron [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.999246] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337284, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.107438] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cc4964-5d6c-4a41-97f5-c0a9d2456c79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.117599] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd93a5a-ffad-4cfd-955c-153802fb92f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.158983] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69fe9ad-3327-46fa-9d53-289de872a70c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.168863] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dc563e-0eac-466e-ad38-65c1fece0c3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.187046] env[61006]: DEBUG nova.compute.provider_tree [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.203314] env[61006]: DEBUG oslo_concurrency.lockutils [req-8bfeb6e4-57d6-4ccc-9e0a-a71bf4b4f3eb req-29bbfabb-8dfc-452b-adf2-1b7bb2d7e873 service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.345855] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.432991] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337282, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.477377] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2d933a0b-41d7-4538-bad7-26aca6f47302 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.305s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.491682] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337284, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66872} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.492164] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1/97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.492654] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.492929] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58a5e607-6955-4b89-be33-6c6643726bef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.501961] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 929.501961] env[61006]: value = "task-1337285" [ 929.501961] env[61006]: _type = "Task" [ 929.501961] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.513924] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.691386] env[61006]: DEBUG nova.scheduler.client.report [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.727148] env[61006]: DEBUG nova.compute.manager [req-13f5840f-26df-4719-aa77-eed93b6f1d5d req-febbadf6-763c-40bd-901e-d64f3627dc0f service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Received event network-vif-deleted-919d334d-6479-4c18-a101-55dab88ea5de {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.727148] env[61006]: INFO nova.compute.manager [req-13f5840f-26df-4719-aa77-eed93b6f1d5d req-febbadf6-763c-40bd-901e-d64f3627dc0f service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Neutron deleted interface 919d334d-6479-4c18-a101-55dab88ea5de; detaching it from the instance and deleting it from the info cache [ 929.727148] env[61006]: DEBUG nova.network.neutron [req-13f5840f-26df-4719-aa77-eed93b6f1d5d req-febbadf6-763c-40bd-901e-d64f3627dc0f service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.859784] env[61006]: DEBUG nova.network.neutron [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.927199] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337282, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.014439] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337285, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082831} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.014652] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.015551] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0051cad4-68de-4917-a71b-2a483745ea93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.036296] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.036690] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.037106] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "370f2153-adca-4513-8549-2bb7499cf913-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.037548] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.037712] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.048879] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1/97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.049837] env[61006]: INFO nova.compute.manager [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Terminating instance [ 930.052040] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0977f3b9-517f-4e7b-bd14-24a8c1fea509 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.070868] env[61006]: DEBUG nova.compute.manager [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.071053] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.072091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b0dfc-498b-43e6-98f4-5bb3546c8cec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.083469] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.085508] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fca04c5-3393-4071-9cf4-bab9767e50c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.087740] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 930.087740] env[61006]: value = "task-1337286" [ 930.087740] env[61006]: _type = "Task" [ 930.087740] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.096761] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 930.096761] env[61006]: value = "task-1337287" [ 930.096761] env[61006]: _type = "Task" [ 930.096761] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.100286] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.109099] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.197716] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.201166] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.855s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.203882] env[61006]: INFO nova.compute.claims [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.223241] env[61006]: INFO nova.scheduler.client.report [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Deleted allocations for instance 1d6ede6a-5601-4da4-baa1-5090a112ec2d [ 930.232505] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47c88160-168d-4fa6-8968-6642408cdc5b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.243496] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b9a68f-000a-45e5-a9c5-875046aa5a61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.286478] env[61006]: DEBUG nova.compute.manager [req-13f5840f-26df-4719-aa77-eed93b6f1d5d req-febbadf6-763c-40bd-901e-d64f3627dc0f service nova] [instance: 2259c42a-79f8-455f-b29f-464705526932] Detach interface failed, port_id=919d334d-6479-4c18-a101-55dab88ea5de, reason: Instance 2259c42a-79f8-455f-b29f-464705526932 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 930.365505] env[61006]: INFO nova.compute.manager [-] [instance: 2259c42a-79f8-455f-b29f-464705526932] Took 1.38 seconds to deallocate network for instance. [ 930.374468] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 930.375669] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd832a3-e6f8-4f3c-8d7f-1a5c04237e8d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.384455] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 930.384653] env[61006]: ERROR oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk due to incomplete transfer. [ 930.384896] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-79493fd5-966d-4df9-a84d-f6f94ca116ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.401815] env[61006]: DEBUG oslo_vmware.rw_handles [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520647e6-a67f-37fc-add3-a1a9666b8cf8/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 930.402102] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Uploaded image 87693d49-a593-43f4-8a3e-5dc9080f7010 to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 930.404450] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 930.404760] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9e7fd60a-fe10-4297-b0d8-ce2b7d90e264 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.413193] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 930.413193] env[61006]: value = "task-1337288" [ 930.413193] env[61006]: _type = "Task" [ 930.413193] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.427678] env[61006]: DEBUG oslo_vmware.api [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337282, 'name': ReconfigVM_Task, 'duration_secs': 1.924242} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.431670] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285375', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'name': 'volume-e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83', 'serial': 'e9dcf69a-c6ec-4cf0-9a48-a98dca7a9c83'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 930.434349] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337288, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.603020] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337286, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.614237] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337287, 'name': PowerOffVM_Task, 'duration_secs': 0.439956} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.614237] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.614237] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.614237] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-886c8bdd-1804-4613-ac94-4033b7008dd2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.734021] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.734021] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.734021] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleting the datastore file [datastore1] 370f2153-adca-4513-8549-2bb7499cf913 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.738082] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-096c88f3-6f39-40f9-9701-6e40e6b3c60d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.739841] env[61006]: DEBUG oslo_concurrency.lockutils [None req-3fdaaca7-9d00-4e83-bec5-dd0c628f623c tempest-ServerTagsTestJSON-2088471430 tempest-ServerTagsTestJSON-2088471430-project-member] Lock "1d6ede6a-5601-4da4-baa1-5090a112ec2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.238s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.747991] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 930.747991] env[61006]: value = "task-1337290" [ 930.747991] env[61006]: _type = "Task" [ 930.747991] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.760340] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.881784] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.926466] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337288, 'name': Destroy_Task, 'duration_secs': 0.507336} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.927144] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Destroyed the VM [ 930.927144] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 930.927292] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6f2e2b70-5600-442a-9b75-4f415c2cf7c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.937487] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 930.937487] env[61006]: value = "task-1337291" [ 930.937487] env[61006]: _type = "Task" [ 930.937487] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.947579] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337291, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.984909] env[61006]: DEBUG nova.objects.instance [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lazy-loading 'flavor' on Instance uuid d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.100368] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337286, 'name': ReconfigVM_Task, 'duration_secs': 0.703415} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.100743] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1/97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.101453] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baadf229-227b-408a-af4b-4b6413a26c32 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.109855] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 931.109855] env[61006]: value = "task-1337292" [ 931.109855] env[61006]: _type = "Task" [ 931.109855] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.124534] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337292, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.260497] env[61006]: DEBUG oslo_vmware.api [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25045} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.260782] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.260966] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.261165] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.261336] env[61006]: INFO nova.compute.manager [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Took 1.19 seconds to destroy the instance on the hypervisor. [ 931.261585] env[61006]: DEBUG oslo.service.loopingcall [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.261785] env[61006]: DEBUG nova.compute.manager [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.262602] env[61006]: DEBUG nova.network.neutron [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.341352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.341598] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.460342] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337291, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.537406] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb90c97-b03e-4479-99ad-f33baf46a2a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.549037] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae72e7b-f075-4ccf-9b89-24b52d74e176 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.593401] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7c78d6-294f-4def-b38b-d550f8c3132c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.605395] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68f5512-3065-4c00-ae95-671d692483d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.620885] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337292, 'name': Rename_Task, 'duration_secs': 0.211207} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.632745] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.632745] env[61006]: DEBUG nova.compute.provider_tree [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.633841] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0dbbd6ae-ebbc-431d-8cb9-8507547b073b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.643749] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 931.643749] env[61006]: value = "task-1337293" [ 931.643749] env[61006]: _type = "Task" [ 931.643749] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.653373] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.848825] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 931.864665] env[61006]: DEBUG nova.compute.manager [req-9867726f-cfec-4a7d-8479-a6ca6b2429f4 req-aaa3f307-6dbc-41cb-bb1b-2464e1b9e700 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Received event network-vif-deleted-f537f6f0-2a6c-4479-84ea-686716df3690 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.865183] env[61006]: INFO nova.compute.manager [req-9867726f-cfec-4a7d-8479-a6ca6b2429f4 req-aaa3f307-6dbc-41cb-bb1b-2464e1b9e700 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Neutron deleted interface f537f6f0-2a6c-4479-84ea-686716df3690; detaching it from the instance and deleting it from the info cache [ 931.865616] env[61006]: DEBUG nova.network.neutron [req-9867726f-cfec-4a7d-8479-a6ca6b2429f4 req-aaa3f307-6dbc-41cb-bb1b-2464e1b9e700 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.920278] env[61006]: DEBUG nova.compute.manager [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.921593] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb195f26-ca0c-41d8-bf20-156b64d8d7f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.954853] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337291, 'name': RemoveSnapshot_Task, 'duration_secs': 0.659071} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.955374] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 931.955843] env[61006]: DEBUG nova.compute.manager [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.957170] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bab2c4-7f4f-4457-b646-9345f7acfb6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.993938] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d4b4b4df-1620-485d-a595-39af98ef48f2 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.885s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.105032] env[61006]: DEBUG nova.network.neutron [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.137059] env[61006]: DEBUG nova.scheduler.client.report [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.156543] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337293, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.369223] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9d58033-8e22-42af-9958-8dac1fde8bea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.382858] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98954e4-6a36-4f77-bdc2-a45f70addc61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.395090] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.431859] env[61006]: DEBUG nova.compute.manager [req-9867726f-cfec-4a7d-8479-a6ca6b2429f4 req-aaa3f307-6dbc-41cb-bb1b-2464e1b9e700 service nova] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Detach interface failed, port_id=f537f6f0-2a6c-4479-84ea-686716df3690, reason: Instance 370f2153-adca-4513-8549-2bb7499cf913 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 932.436754] env[61006]: INFO nova.compute.manager [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] instance snapshotting [ 932.441073] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb10c6c-1391-429b-bf8f-df6504c426c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.481458] env[61006]: INFO nova.compute.manager [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Shelve offloading [ 932.484025] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7ea25d-2d5a-4536-96e3-78e6d4cbe5b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.488022] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.488667] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a31e72c-9c2b-4f52-8bc6-105925c9b720 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.501247] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 932.501247] env[61006]: value = "task-1337294" [ 932.501247] env[61006]: _type = "Task" [ 932.501247] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.513038] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 932.514117] env[61006]: DEBUG nova.compute.manager [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.514727] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b03220-c550-4e98-a1a6-c6cc108d034b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.522161] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.522547] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.522547] env[61006]: DEBUG nova.network.neutron [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.607574] env[61006]: INFO nova.compute.manager [-] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Took 1.35 seconds to deallocate network for instance. [ 932.643186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.643574] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.649205] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.768s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.649363] env[61006]: DEBUG nova.objects.instance [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lazy-loading 'resources' on Instance uuid 2259c42a-79f8-455f-b29f-464705526932 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.664622] env[61006]: DEBUG oslo_vmware.api [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337293, 'name': PowerOnVM_Task, 'duration_secs': 0.680291} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.665886] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.666127] env[61006]: INFO nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Took 10.70 seconds to spawn the instance on the hypervisor. [ 932.666329] env[61006]: DEBUG nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.667231] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b38012-8993-42bc-bdc1-2686b571142e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.900306] env[61006]: DEBUG nova.compute.manager [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.902922] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d36c32-736c-4603-82b7-aa67f7545fe4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.996733] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 932.997036] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3b5bc1f6-8901-463e-afcf-683d0c8bf663 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.006164] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 933.006164] env[61006]: value = "task-1337295" [ 933.006164] env[61006]: _type = "Task" [ 933.006164] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.019174] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337295, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.048477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.048477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.114726] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.152932] env[61006]: DEBUG nova.compute.utils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.154938] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.155129] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.199373] env[61006]: INFO nova.compute.manager [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Took 26.67 seconds to build instance. [ 933.273738] env[61006]: DEBUG nova.policy [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.417502] env[61006]: INFO nova.compute.manager [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] instance snapshotting [ 933.422252] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68dde27-0c61-4111-b5f0-838029b22958 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.448919] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc1c2a9-4bf6-414c-a589-a332e22de061 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.506866] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50bddc6-2776-4ef4-afe6-51c5383de576 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.524441] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b30c4d7-23af-4b10-9c66-c718cf3cf2d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.529082] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337295, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.573224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.573376] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.575262] env[61006]: INFO nova.compute.manager [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Detaching volume 9f7e0232-99d9-446e-aca8-77c70e411002 [ 933.578822] env[61006]: DEBUG nova.network.neutron [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.582380] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ddf08c-8ce2-487e-b06e-dba05b65ad43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.605173] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e1b02d-33f7-4ed8-ac51-a0b3df2dd889 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.627902] env[61006]: DEBUG nova.compute.provider_tree [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.629968] env[61006]: INFO nova.virt.block_device [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Attempting to driver detach volume 9f7e0232-99d9-446e-aca8-77c70e411002 from mountpoint /dev/sdc [ 933.630220] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 933.630431] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285384', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'name': 'volume-9f7e0232-99d9-446e-aca8-77c70e411002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'serial': '9f7e0232-99d9-446e-aca8-77c70e411002'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 933.631682] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b83f696-bab6-4615-ab54-c2ffefbf47bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.660239] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.665850] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc3d2b7-5e6c-4133-9440-3bedf7ac67a9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.679620] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3084916d-ae92-45f5-92fc-60946b59d121 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.703057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bafc6191-d437-46a8-9a7e-2151786123bc tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.190s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.703974] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cf50d2-8170-4fda-9880-bfe2e4e3f63e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.722437] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] The volume has not been displaced from its original location: [datastore2] volume-9f7e0232-99d9-446e-aca8-77c70e411002/volume-9f7e0232-99d9-446e-aca8-77c70e411002.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 933.727950] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfiguring VM instance instance-0000004c to detach disk 2002 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.729328] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-398b3d02-a7eb-4d52-996a-e9c0aaa71ffe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.750459] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 933.750459] env[61006]: value = "task-1337296" [ 933.750459] env[61006]: _type = "Task" [ 933.750459] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.759606] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337296, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.913351] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Successfully created port: 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.959997] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.960404] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.960708] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.960946] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.962145] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.967025] env[61006]: INFO nova.compute.manager [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Terminating instance [ 933.969133] env[61006]: DEBUG nova.compute.manager [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.969393] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.970504] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 933.971340] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dabd2c8-d916-4fed-80ec-8695421165e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.974569] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e14f85b1-b9e4-4e02-a152-c25f5c69b2a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.982893] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.984385] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c0ca429-7339-462a-872f-3e2875897268 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.986250] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 933.986250] env[61006]: value = "task-1337297" [ 933.986250] env[61006]: _type = "Task" [ 933.986250] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.992699] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 933.992699] env[61006]: value = "task-1337298" [ 933.992699] env[61006]: _type = "Task" [ 933.992699] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.996167] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337297, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.005655] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337298, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.023038] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337295, 'name': CreateSnapshot_Task, 'duration_secs': 0.848406} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.023038] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 934.023038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11acb1a9-a24d-43f6-8438-00ba77b5d900 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.084286] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.087443] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.139726] env[61006]: DEBUG nova.scheduler.client.report [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.261457] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337296, 'name': ReconfigVM_Task, 'duration_secs': 0.342114} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.261806] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Reconfigured VM instance instance-0000004c to detach disk 2002 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 934.266706] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78ed9ea0-8922-42d8-bc86-c0aba2920520 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.283428] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 934.283428] env[61006]: value = "task-1337299" [ 934.283428] env[61006]: _type = "Task" [ 934.283428] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.293362] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337299, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.497720] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337297, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.507560] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337298, 'name': PowerOffVM_Task, 'duration_secs': 0.197336} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.508047] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.510395] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.510395] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e7c58e6-e350-4908-8176-9602b5cc7437 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.528393] env[61006]: DEBUG nova.compute.manager [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-vif-unplugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.528393] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.528393] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.528393] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.528393] env[61006]: DEBUG nova.compute.manager [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] No waiting events found dispatching network-vif-unplugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.528393] env[61006]: WARNING nova.compute.manager [req-0bd137c0-9ecd-42f6-ae8b-382259299446 req-2d2e2464-831e-4c00-9c76-42c4ddbda75d service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received unexpected event network-vif-unplugged-9230a732-200c-4084-8c6c-a5892e9a50ba for instance with vm_state shelved and task_state shelving_offloading. [ 934.545486] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 934.546382] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9c1c4c00-cae3-4de5-a970-d70c2a640ac3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.556720] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 934.556720] env[61006]: value = "task-1337301" [ 934.556720] env[61006]: _type = "Task" [ 934.556720] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.569997] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337301, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.583746] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.584211] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.584528] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore1] 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.585666] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3737e5a8-d8d3-44b8-92fc-459f520cd746 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.596894] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 934.596894] env[61006]: value = "task-1337302" [ 934.596894] env[61006]: _type = "Task" [ 934.596894] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.610238] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.614949] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.645742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.996s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.650059] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.255s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.652866] env[61006]: INFO nova.compute.claims [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.670893] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.674555] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.676735] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5da185c-820a-49ed-9539-84af6d76f245 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.684045] env[61006]: INFO nova.scheduler.client.report [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance 2259c42a-79f8-455f-b29f-464705526932 [ 934.703278] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.705694] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19312ac0-2b6f-47a4-a559-977494ba6526 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.717215] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.717592] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.717592] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.717807] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.717900] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.718109] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.718339] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.718580] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.718694] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.718863] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.719079] env[61006]: DEBUG nova.virt.hardware [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.720490] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f163b82d-3875-4a7e-8420-d464b15b9c14 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.731859] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc17e8e-1ee4-47eb-9243-ba65db17661a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.796586] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337299, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.798395] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.798766] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.798966] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.799323] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71df0207-2bef-4994-86af-66ba887a1664 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.807959] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 934.807959] env[61006]: value = "task-1337304" [ 934.807959] env[61006]: _type = "Task" [ 934.807959] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.820369] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.998456] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337297, 'name': CreateSnapshot_Task, 'duration_secs': 0.755616} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.998795] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 934.999622] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28389a84-3647-4dbe-a34d-1a7d5e654f70 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.073083] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337301, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.108972] env[61006]: DEBUG oslo_vmware.api [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224329} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.108972] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.108972] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.109205] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.109416] env[61006]: INFO nova.compute.manager [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 935.109676] env[61006]: DEBUG oslo.service.loopingcall [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.109926] env[61006]: DEBUG nova.compute.manager [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.110089] env[61006]: DEBUG nova.network.neutron [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.169688] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.170026] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.201271] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e27bb822-d51b-4d95-829e-bffeedac45d4 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "2259c42a-79f8-455f-b29f-464705526932" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.920s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.294837] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337299, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.317716] env[61006]: DEBUG oslo_vmware.api [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.347114} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.317988] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.318210] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.318396] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.339633] env[61006]: INFO nova.scheduler.client.report [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted allocations for instance 66c93148-b91a-4d22-84af-f410c8e10875 [ 935.498019] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 935.498979] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edae3d8-532e-4ed3-864f-aa6a937e370c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.506295] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 935.506438] env[61006]: ERROR oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk due to incomplete transfer. [ 935.506674] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-75c394cc-de60-439a-b3d9-f05c35296246 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.520628] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 935.521528] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-59e5ec4a-d731-4c69-ab54-cfca1d0961e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.525580] env[61006]: DEBUG oslo_vmware.rw_handles [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a5a343-92ea-2041-167b-81bc8ae69f4f/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 935.525816] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Uploaded image 767836f1-ca1e-4962-9506-b705f9ab4d4c to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 935.528279] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 935.529090] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f670b5e6-a55a-4ac5-8f59-97926927ed9c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.537868] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 935.537868] env[61006]: value = "task-1337306" [ 935.537868] env[61006]: _type = "Task" [ 935.537868] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.539229] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 935.539229] env[61006]: value = "task-1337305" [ 935.539229] env[61006]: _type = "Task" [ 935.539229] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.553345] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337305, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.557339] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337306, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.576542] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337301, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.663914] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Successfully updated port: 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.674224] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.797248] env[61006]: DEBUG oslo_vmware.api [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337299, 'name': ReconfigVM_Task, 'duration_secs': 1.132386} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.797944] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285384', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'name': 'volume-9f7e0232-99d9-446e-aca8-77c70e411002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd174a8ec-867e-4fea-b878-2a9af1476949', 'attached_at': '', 'detached_at': '', 'volume_id': '9f7e0232-99d9-446e-aca8-77c70e411002', 'serial': '9f7e0232-99d9-446e-aca8-77c70e411002'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 935.849501] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.909648] env[61006]: DEBUG nova.network.neutron [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.927654] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd732cf7-4485-4e8f-a2de-a9e1e0324555 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.937038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4b6481-5221-434e-8846-0ae8bd6f0694 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.970019] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37ae533-f1d6-4eee-9c73-07f691d71a11 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.979349] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078e2982-0f03-4055-b1f9-6c2b9ff56647 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.994529] env[61006]: DEBUG nova.compute.provider_tree [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.053433] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337306, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.056745] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337305, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.075730] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337301, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.171260] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.171467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.171713] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.196099] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.375723] env[61006]: DEBUG nova.objects.instance [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lazy-loading 'flavor' on Instance uuid d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.413032] env[61006]: INFO nova.compute.manager [-] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Took 1.30 seconds to deallocate network for instance. [ 936.497599] env[61006]: DEBUG nova.scheduler.client.report [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.559154] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337306, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.563842] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337305, 'name': Destroy_Task, 'duration_secs': 0.779937} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.565847] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.566077] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing instance network info cache due to event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.566318] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.566473] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.566659] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 936.567925] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroyed the VM [ 936.568201] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 936.572897] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fd1f4a6d-61fe-4ab6-8d02-4e85dd647a29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.583227] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337301, 'name': CloneVM_Task, 'duration_secs': 1.98695} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.584677] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Created linked-clone VM from snapshot [ 936.585053] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 936.585053] env[61006]: value = "task-1337307" [ 936.585053] env[61006]: _type = "Task" [ 936.585053] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.585760] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bb71c7-fab1-4e9a-9a98-971680ea33d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.598008] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337307, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.601332] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Uploading image 0ad8cc0d-bf01-410f-a9ad-04138005b41e {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 936.612366] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 936.612669] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-afda3a32-692d-49cb-9cd3-68f500893fe4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.623944] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 936.623944] env[61006]: value = "task-1337308" [ 936.623944] env[61006]: _type = "Task" [ 936.623944] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.631610] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337308, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.713314] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.713590] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.760893] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.920065] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.973118] env[61006]: DEBUG nova.network.neutron [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.004957] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.005738] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.009060] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.894s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.009391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.013313] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.397s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.016155] env[61006]: INFO nova.compute.claims [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.040333] env[61006]: INFO nova.scheduler.client.report [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocations for instance 370f2153-adca-4513-8549-2bb7499cf913 [ 937.054792] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337306, 'name': CloneVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.100606] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337307, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.134571] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337308, 'name': Destroy_Task, 'duration_secs': 0.377519} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.134871] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Destroyed the VM [ 937.135162] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 937.135440] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-118f0ef1-9908-4dbf-9087-74667b10fe9b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.142926] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 937.142926] env[61006]: value = "task-1337309" [ 937.142926] env[61006]: _type = "Task" [ 937.142926] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.153105] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337309, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.216531] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.327616] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updated VIF entry in instance network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.327989] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9230a732-20", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.383527] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8b86c75f-7469-4714-b972-ebf14578e73f tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.336s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.476166] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.476494] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Instance network_info: |[{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.476964] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:f3:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d945d09-25a9-4087-96d5-f0be90bcddff', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.486750] env[61006]: DEBUG oslo.service.loopingcall [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.487060] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.487345] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d1cc99b-31ab-4416-b02b-3794b6fd1add {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.513335] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.513335] env[61006]: value = "task-1337310" [ 937.513335] env[61006]: _type = "Task" [ 937.513335] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.524106] env[61006]: DEBUG nova.compute.utils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.527764] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337310, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.528555] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.528738] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.554937] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337306, 'name': CloneVM_Task, 'duration_secs': 1.524444} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.555488] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37899274-02d1-4e2b-99fa-c0a993c14d6b tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "370f2153-adca-4513-8549-2bb7499cf913" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.519s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.556735] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Created linked-clone VM from snapshot [ 937.557530] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0807a5a-4ced-495b-899b-83d359137828 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.566315] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Uploading image 00666797-346f-4a13-84a1-8a64b7da3432 {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 937.576769] env[61006]: DEBUG nova.policy [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cca6e2806aa45208ae618f6a78ccc0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe9eabeec9a941e68a9eae559e24ff4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.598305] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 937.598305] env[61006]: value = "vm-285393" [ 937.598305] env[61006]: _type = "VirtualMachine" [ 937.598305] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 937.598607] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ba9ba5b3-6a3c-42ec-a0f5-8c9ae2c54e6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.606540] env[61006]: DEBUG oslo_vmware.api [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337307, 'name': RemoveSnapshot_Task, 'duration_secs': 0.612856} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.607817] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 937.608084] env[61006]: INFO nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 15.84 seconds to snapshot the instance on the hypervisor. [ 937.610524] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lease: (returnval){ [ 937.610524] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b7161c-2165-3baa-cbaf-efe3fe27e710" [ 937.610524] env[61006]: _type = "HttpNfcLease" [ 937.610524] env[61006]: } obtained for exporting VM: (result){ [ 937.610524] env[61006]: value = "vm-285393" [ 937.610524] env[61006]: _type = "VirtualMachine" [ 937.610524] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 937.610852] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the lease: (returnval){ [ 937.610852] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b7161c-2165-3baa-cbaf-efe3fe27e710" [ 937.610852] env[61006]: _type = "HttpNfcLease" [ 937.610852] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 937.619162] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 937.619162] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b7161c-2165-3baa-cbaf-efe3fe27e710" [ 937.619162] env[61006]: _type = "HttpNfcLease" [ 937.619162] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 937.655953] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337309, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.738982] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.830694] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.831118] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Received event network-vif-deleted-23bdf050-f167-4c83-a7c5-8b64df012cd0 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.831188] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-vif-plugged-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.831383] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.831617] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.831788] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.831958] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] No waiting events found dispatching network-vif-plugged-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.832185] env[61006]: WARNING nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received unexpected event network-vif-plugged-5d945d09-25a9-4087-96d5-f0be90bcddff for instance with vm_state building and task_state spawning. [ 937.832372] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.832534] env[61006]: DEBUG nova.compute.manager [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing instance network info cache due to event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.832931] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.832931] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.833073] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.866623] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Successfully created port: de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.876308] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.027451] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337310, 'name': CreateVM_Task, 'duration_secs': 0.386427} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.027607] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.028255] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.028427] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.028750] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.029297] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.031771] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42474ddd-59a5-4eed-b2e7-f825115dedba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.040180] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 938.040180] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527b5fc8-6dfa-0df7-f4f0-988d68d5d36c" [ 938.040180] env[61006]: _type = "Task" [ 938.040180] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.049139] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527b5fc8-6dfa-0df7-f4f0-988d68d5d36c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.129633] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 938.129633] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b7161c-2165-3baa-cbaf-efe3fe27e710" [ 938.129633] env[61006]: _type = "HttpNfcLease" [ 938.129633] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 938.131353] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 938.131353] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b7161c-2165-3baa-cbaf-efe3fe27e710" [ 938.131353] env[61006]: _type = "HttpNfcLease" [ 938.131353] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 938.131353] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0993ec6-c974-4fd5-ac45-c85240618506 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.140783] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 938.140783] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 938.227662] env[61006]: DEBUG nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Found 3 images (rotation: 2) {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 938.227855] env[61006]: DEBUG nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Rotating out 1 backups {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 938.228036] env[61006]: DEBUG nova.compute.manager [None req-857da68c-5d16-49a4-999b-d14d56e50188 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleting image e72dabda-ebc9-4e91-a9b2-f12e7a212e79 {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 938.237027] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337309, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.250875] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5105e1dc-833e-4297-9d43-29b63cb8d2aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.415917] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7aa9f8c-e57e-4943-a056-9f42269e9b29 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.424607] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5792ff7-bd20-44f0-b6b1-4888cdba7851 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.462366] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178058b5-2734-4dff-b8bb-053fa7a0d232 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.471455] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5a0616-b729-4295-8829-b7d045408114 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.487743] env[61006]: DEBUG nova.compute.provider_tree [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.552286] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527b5fc8-6dfa-0df7-f4f0-988d68d5d36c, 'name': SearchDatastore_Task, 'duration_secs': 0.029941} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.552286] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.552286] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.552575] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.552654] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.553318] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.553612] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6e903e5-1c65-4769-a2ca-0714b09093cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.564478] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.564766] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.565600] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7614e315-757f-4448-8645-51bed5e45be4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.573143] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 938.573143] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52433980-ad99-62c2-04e5-1e48039cce38" [ 938.573143] env[61006]: _type = "Task" [ 938.573143] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.583218] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52433980-ad99-62c2-04e5-1e48039cce38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.636192] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updated VIF entry in instance network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.636380] env[61006]: DEBUG nova.network.neutron [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.721732] env[61006]: DEBUG oslo_vmware.api [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337309, 'name': RemoveSnapshot_Task, 'duration_secs': 1.104115} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.722084] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 938.866224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.866836] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.867344] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.867605] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.867788] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.870494] env[61006]: INFO nova.compute.manager [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Terminating instance [ 938.873464] env[61006]: DEBUG nova.compute.manager [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.873672] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.874609] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36bd44bc-b280-487e-a8d8-95d4d7ef749d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.883235] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.883596] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fd8f8fb-19c7-46d2-a7c6-6e7d7bc27133 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.892042] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 938.892042] env[61006]: value = "task-1337312" [ 938.892042] env[61006]: _type = "Task" [ 938.892042] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.902195] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337312, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.992244] env[61006]: DEBUG nova.scheduler.client.report [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.042638] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.075559] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.075889] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.076127] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.076330] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.076493] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.076775] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.076912] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.077105] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.077330] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.077512] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.077769] env[61006]: DEBUG nova.virt.hardware [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.080354] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df349a5-1430-43d8-a390-a866eea3e02b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.095378] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117beb6b-49f7-4b29-831c-42f489620d3d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.100318] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52433980-ad99-62c2-04e5-1e48039cce38, 'name': SearchDatastore_Task, 'duration_secs': 0.012675} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.101728] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a95d5ab0-0677-4bc8-b6bc-b699594a1763 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.115064] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 939.115064] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529553b5-cbb6-2da2-9d58-a2d9c18b9136" [ 939.115064] env[61006]: _type = "Task" [ 939.115064] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.124296] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529553b5-cbb6-2da2-9d58-a2d9c18b9136, 'name': SearchDatastore_Task, 'duration_secs': 0.011957} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.124934] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.126138] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] f048bd9c-048b-4439-bbbf-dfcee7f18d84/f048bd9c-048b-4439-bbbf-dfcee7f18d84.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.126138] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb0fc935-0a14-4a31-a251-a18a880f98ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.135064] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 939.135064] env[61006]: value = "task-1337313" [ 939.135064] env[61006]: _type = "Task" [ 939.135064] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.139464] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce1cfce6-b6a3-42d4-99d5-0bd1fc915fd6 req-21d623c0-390d-4efa-875b-add597580126 service nova] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.145830] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.227911] env[61006]: WARNING nova.compute.manager [None req-936f78c2-bd77-4732-96ba-936531f07c48 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Image not found during snapshot: nova.exception.ImageNotFound: Image 0ad8cc0d-bf01-410f-a9ad-04138005b41e could not be found. [ 939.412454] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337312, 'name': PowerOffVM_Task, 'duration_secs': 0.216673} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.413080] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.413295] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.413703] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2c17dab-8f7e-4804-809b-0215b7e79816 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.498317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.501055] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.507870] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.656s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.508835] env[61006]: DEBUG nova.objects.instance [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'resources' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.522681] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.523263] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.523536] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Deleting the datastore file [datastore1] d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.523890] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-186b9b3d-b9a7-4ebe-9394-28fc9806ab71 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.534715] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for the task: (returnval){ [ 939.534715] env[61006]: value = "task-1337315" [ 939.534715] env[61006]: _type = "Task" [ 939.534715] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.546598] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.644287] env[61006]: DEBUG nova.compute.manager [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Received event network-vif-plugged-de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.644724] env[61006]: DEBUG oslo_concurrency.lockutils [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.644999] env[61006]: DEBUG oslo_concurrency.lockutils [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.645246] env[61006]: DEBUG oslo_concurrency.lockutils [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.645643] env[61006]: DEBUG nova.compute.manager [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] No waiting events found dispatching network-vif-plugged-de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.645862] env[61006]: WARNING nova.compute.manager [req-3b0f21d5-07fc-4ae9-a410-4e82abb6622e req-d79ea691-86ba-46f0-887a-de653d736e39 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Received unexpected event network-vif-plugged-de9a3378-0e2c-485b-842b-d44c829bee0f for instance with vm_state building and task_state spawning. [ 939.650918] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337313, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.789549] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Successfully updated port: de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.976628] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "40500a0d-33f3-492a-9573-ed10e45642c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.976971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.977303] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.977616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.977799] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.980511] env[61006]: INFO nova.compute.manager [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Terminating instance [ 939.982618] env[61006]: DEBUG nova.compute.manager [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.982873] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.983883] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeea32bd-50ca-4b4e-bda4-19a2145c72d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.992773] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.993448] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0aca346b-6954-423e-a377-a6fd804b04e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.001159] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 940.001159] env[61006]: value = "task-1337316" [ 940.001159] env[61006]: _type = "Task" [ 940.001159] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.015099] env[61006]: DEBUG nova.compute.utils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.016996] env[61006]: DEBUG nova.objects.instance [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'numa_topology' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.018453] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337316, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.019279] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.019887] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.049938] env[61006]: DEBUG oslo_vmware.api [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Task: {'id': task-1337315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273555} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.050169] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.050741] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.051043] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.051311] env[61006]: INFO nova.compute.manager [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Took 1.18 seconds to destroy the instance on the hypervisor. [ 940.051641] env[61006]: DEBUG oslo.service.loopingcall [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.052154] env[61006]: DEBUG nova.compute.manager [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.052261] env[61006]: DEBUG nova.network.neutron [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 940.082991] env[61006]: DEBUG nova.policy [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7c04e3804464cdc8c2925f7bbcf5369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4e87e7cb3f2444a8cc7bb7e65a99757', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.149317] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580125} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.150489] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] f048bd9c-048b-4439-bbbf-dfcee7f18d84/f048bd9c-048b-4439-bbbf-dfcee7f18d84.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.150806] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.151132] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-933bda22-b8cc-414c-b0af-94ce29f2d359 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.162250] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 940.162250] env[61006]: value = "task-1337317" [ 940.162250] env[61006]: _type = "Task" [ 940.162250] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.177797] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.295560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.295752] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.295934] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.512868] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337316, 'name': PowerOffVM_Task, 'duration_secs': 0.214774} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.513418] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.513711] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.514135] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc00cd91-e636-4c06-8d10-611e3beae99c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.519818] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.522885] env[61006]: DEBUG nova.objects.base [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Object Instance<66c93148-b91a-4d22-84af-f410c8e10875> lazy-loaded attributes: resources,numa_topology {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 940.592825] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.593811] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.596021] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleting the datastore file [datastore2] 40500a0d-33f3-492a-9573-ed10e45642c5 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.596021] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-91361cc5-9c83-4f01-83a9-a9be73c6f808 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.611679] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for the task: (returnval){ [ 940.611679] env[61006]: value = "task-1337319" [ 940.611679] env[61006]: _type = "Task" [ 940.611679] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.624783] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.631778] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.632114] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.633074] env[61006]: DEBUG nova.compute.manager [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.634048] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea070eb-77f0-4ed1-9316-187455618b5e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.649508] env[61006]: DEBUG nova.compute.manager [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61006) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 940.650568] env[61006]: DEBUG nova.objects.instance [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.686357] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082468} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.687814] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.688321] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d01202-7ddd-493c-857c-0ce4a90181fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.731192] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] f048bd9c-048b-4439-bbbf-dfcee7f18d84/f048bd9c-048b-4439-bbbf-dfcee7f18d84.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.735437] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3863f336-c463-4a96-be39-12adc22523ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.761443] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Successfully created port: aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.776531] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 940.776531] env[61006]: value = "task-1337320" [ 940.776531] env[61006]: _type = "Task" [ 940.776531] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.791023] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.890056] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.921731] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2227416-eb66-4bda-9eae-5fdc76f68280 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.931495] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123b9647-d015-4c7e-adaf-1d79f40fcc4d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.966488] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8043a7-963a-4ed2-a563-6d1ea8f76124 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.977404] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0f59a1-14c1-4440-a8b9-f63f7bc68bef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.999025] env[61006]: DEBUG nova.compute.provider_tree [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.119221] env[61006]: DEBUG nova.network.neutron [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.127971] env[61006]: DEBUG oslo_vmware.api [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Task: {'id': task-1337319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317799} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.128543] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.128832] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.129059] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.129248] env[61006]: INFO nova.compute.manager [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 941.129612] env[61006]: DEBUG oslo.service.loopingcall [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.131044] env[61006]: DEBUG nova.compute.manager [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.131178] env[61006]: DEBUG nova.network.neutron [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.158678] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.159097] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2656c8db-b399-4c1e-b9c7-bf1a048a7f07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.170711] env[61006]: DEBUG oslo_vmware.api [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 941.170711] env[61006]: value = "task-1337321" [ 941.170711] env[61006]: _type = "Task" [ 941.170711] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.182037] env[61006]: DEBUG oslo_vmware.api [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.296320] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337320, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.498858] env[61006]: DEBUG nova.scheduler.client.report [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.531830] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.558544] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.558826] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.558981] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.559272] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.559429] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.559581] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.559790] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.559950] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.560346] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.560547] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.560816] env[61006]: DEBUG nova.virt.hardware [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.562008] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c585bd77-8621-40bc-adf8-6ed7c2d0e47f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.573419] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa590525-1f8c-45ce-91ce-d5cca042fa73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.610541] env[61006]: DEBUG nova.compute.manager [req-62b3b7d7-37ff-4fac-90ca-cd47416838ce req-3b387da1-e867-41ec-8c67-d996fddf06b7 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Received event network-vif-deleted-7fe1873c-f1dd-498a-893b-b179d4714d5d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.610869] env[61006]: INFO nova.compute.manager [req-62b3b7d7-37ff-4fac-90ca-cd47416838ce req-3b387da1-e867-41ec-8c67-d996fddf06b7 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Neutron deleted interface 7fe1873c-f1dd-498a-893b-b179d4714d5d; detaching it from the instance and deleting it from the info cache [ 941.611179] env[61006]: DEBUG nova.network.neutron [req-62b3b7d7-37ff-4fac-90ca-cd47416838ce req-3b387da1-e867-41ec-8c67-d996fddf06b7 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.621027] env[61006]: DEBUG nova.network.neutron [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.623208] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.623630] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Instance network_info: |[{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.625147] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:e8:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb68953b-dee5-4d9d-b47b-277336ba76dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de9a3378-0e2c-485b-842b-d44c829bee0f', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.638242] env[61006]: DEBUG oslo.service.loopingcall [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.639781] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.640155] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47f29844-bd75-4393-8b90-df1a9b40020f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.681060] env[61006]: DEBUG nova.compute.manager [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Received event network-changed-de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.681382] env[61006]: DEBUG nova.compute.manager [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Refreshing instance network info cache due to event network-changed-de9a3378-0e2c-485b-842b-d44c829bee0f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.681718] env[61006]: DEBUG oslo_concurrency.lockutils [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] Acquiring lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.681959] env[61006]: DEBUG oslo_concurrency.lockutils [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] Acquired lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.682254] env[61006]: DEBUG nova.network.neutron [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Refreshing network info cache for port de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.684793] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.684793] env[61006]: value = "task-1337322" [ 941.684793] env[61006]: _type = "Task" [ 941.684793] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.696760] env[61006]: DEBUG oslo_vmware.api [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337321, 'name': PowerOffVM_Task, 'duration_secs': 0.277699} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.697810] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.698354] env[61006]: DEBUG nova.compute.manager [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.700439] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf96fc8-dd5d-4a59-aee2-2625cfa65a72 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.710213] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337322, 'name': CreateVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.788197] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337320, 'name': ReconfigVM_Task, 'duration_secs': 0.63204} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.788491] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfigured VM instance instance-0000005c to attach disk [datastore2] f048bd9c-048b-4439-bbbf-dfcee7f18d84/f048bd9c-048b-4439-bbbf-dfcee7f18d84.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.789138] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05f8b30c-6bd1-47f1-bbc1-1a06036442bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.797317] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 941.797317] env[61006]: value = "task-1337323" [ 941.797317] env[61006]: _type = "Task" [ 941.797317] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.807289] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337323, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.009698] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.505s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.013341] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.817s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.014267] env[61006]: INFO nova.compute.claims [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.038441] env[61006]: DEBUG nova.network.neutron [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.115245] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48d93a7e-18f8-4097-a302-be423e2a6cf8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.127204] env[61006]: INFO nova.compute.manager [-] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Took 2.07 seconds to deallocate network for instance. [ 942.134326] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a2dae9-1c57-44ae-823a-5757511c1b9e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.176806] env[61006]: DEBUG nova.compute.manager [req-62b3b7d7-37ff-4fac-90ca-cd47416838ce req-3b387da1-e867-41ec-8c67-d996fddf06b7 service nova] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Detach interface failed, port_id=7fe1873c-f1dd-498a-893b-b179d4714d5d, reason: Instance 40500a0d-33f3-492a-9573-ed10e45642c5 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 942.200684] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337322, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.227242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5f9554a4-553a-4951-8a54-f383a26fdf09 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.595s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.309539] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337323, 'name': Rename_Task, 'duration_secs': 0.326901} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.309843] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.310115] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f428897-7a3a-4d7d-80de-eb497f686bce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.318323] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 942.318323] env[61006]: value = "task-1337324" [ 942.318323] env[61006]: _type = "Task" [ 942.318323] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.333780] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.510239] env[61006]: DEBUG nova.network.neutron [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updated VIF entry in instance network info cache for port de9a3378-0e2c-485b-842b-d44c829bee0f. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.510239] env[61006]: DEBUG nova.network.neutron [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.527658] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0556b8f6-09af-4fa8-8bc7-1c2ca3be71f3 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.120s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.528257] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.652s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.528610] env[61006]: INFO nova.compute.manager [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Unshelving [ 942.544392] env[61006]: INFO nova.compute.manager [-] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Took 1.41 seconds to deallocate network for instance. [ 942.658547] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.671798] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Successfully updated port: aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.701459] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337322, 'name': CreateVM_Task, 'duration_secs': 0.599828} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.701459] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.702705] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.702874] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.703447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.703564] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c68d85b7-92c6-4e6b-a787-e20de0a0cca2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.709228] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 942.709228] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527666fa-e6c0-9f28-c1c4-a40d8e3c3745" [ 942.709228] env[61006]: _type = "Task" [ 942.709228] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.718252] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527666fa-e6c0-9f28-c1c4-a40d8e3c3745, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.829871] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.012798] env[61006]: DEBUG oslo_concurrency.lockutils [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] Releasing lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.013106] env[61006]: DEBUG nova.compute.manager [req-a7e2797a-af70-4378-a8d1-1c2f33ca07ba req-a2a9c0a7-6cf7-4165-9022-1f44a9b12952 service nova] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Received event network-vif-deleted-0e26f4ec-25ca-4a63-8c6a-b6767eda7456 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.054596] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.173588] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.173735] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.173888] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.220525] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527666fa-e6c0-9f28-c1c4-a40d8e3c3745, 'name': SearchDatastore_Task, 'duration_secs': 0.01367} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.223288] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.223530] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.223764] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.223911] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.224149] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.224617] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5306882-d6eb-490b-8483-0202283523eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.235827] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.236042] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.238155] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8d37c2e-7746-4a48-836a-9ba623ed85b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.241426] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86743f80-1599-40ba-afc5-8dbb0532ee10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.248696] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 943.248696] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]523338fd-6e63-b0fb-f044-5dbaee2fde0b" [ 943.248696] env[61006]: _type = "Task" [ 943.248696] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.254663] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d29504-2670-45b4-8b59-15ea3c0b6a67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.265890] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]523338fd-6e63-b0fb-f044-5dbaee2fde0b, 'name': SearchDatastore_Task, 'duration_secs': 0.011167} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.291436] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-051873dd-8e2e-4cdc-958b-2131fd4f7411 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.294701] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0449f06d-832b-43e9-8b28-e529d681a824 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.303391] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 943.303391] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521a4504-2133-866f-3d60-dd7e89dbeccc" [ 943.303391] env[61006]: _type = "Task" [ 943.303391] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.310908] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a8b443-e7b4-4536-a35e-d04c401f5dfe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.322016] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521a4504-2133-866f-3d60-dd7e89dbeccc, 'name': SearchDatastore_Task, 'duration_secs': 0.01196} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.333472] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.333472] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.337024] env[61006]: DEBUG nova.compute.provider_tree [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.337024] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07f735cb-10d7-406a-9d3f-c0c23a5efe2c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.344560] env[61006]: DEBUG oslo_vmware.api [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337324, 'name': PowerOnVM_Task, 'duration_secs': 1.002478} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.346166] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.346403] env[61006]: INFO nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Took 8.68 seconds to spawn the instance on the hypervisor. [ 943.346597] env[61006]: DEBUG nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.347614] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 943.347614] env[61006]: value = "task-1337325" [ 943.347614] env[61006]: _type = "Task" [ 943.347614] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.348172] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4992ee1-2658-4f60-b56f-b5639aa7c544 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.365551] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.539060] env[61006]: DEBUG nova.compute.utils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.713773] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.722542] env[61006]: DEBUG nova.compute.manager [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Received event network-vif-plugged-aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.722763] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.722976] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Lock "02408ca9-c580-444a-9608-a752146ca499-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.723171] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Lock "02408ca9-c580-444a-9608-a752146ca499-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.723344] env[61006]: DEBUG nova.compute.manager [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] No waiting events found dispatching network-vif-plugged-aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.723512] env[61006]: WARNING nova.compute.manager [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Received unexpected event network-vif-plugged-aa62774d-28ae-41d6-8bbb-3dae4be79437 for instance with vm_state building and task_state spawning. [ 943.723674] env[61006]: DEBUG nova.compute.manager [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Received event network-changed-aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.723828] env[61006]: DEBUG nova.compute.manager [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Refreshing instance network info cache due to event network-changed-aa62774d-28ae-41d6-8bbb-3dae4be79437. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.724645] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Acquiring lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.839737] env[61006]: DEBUG nova.scheduler.client.report [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.865692] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337325, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.877844] env[61006]: INFO nova.compute.manager [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Took 14.55 seconds to build instance. [ 943.896755] env[61006]: DEBUG nova.network.neutron [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [{"id": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "address": "fa:16:3e:b6:01:c1", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa62774d-28", "ovs_interfaceid": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.927291] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.927564] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.996729] env[61006]: DEBUG nova.compute.manager [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Stashing vm_state: stopped {{(pid=61006) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 944.041726] env[61006]: INFO nova.virt.block_device [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Booting with volume ce8406c1-edd0-4c87-b101-c9fc03c9897a at /dev/sdb [ 944.079027] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6649328e-0d0f-4e27-8179-3b4dd60586c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.089597] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1827ac-bf7d-47e5-b9e9-520c27cab002 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.121985] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff73d108-155e-4242-bc52-4553d737ea21 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.132047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03ed1e3-9eeb-4647-8085-a5e1b7e76ec5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.164454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5a38f1-803f-4d2a-856c-013a8de8e583 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.172144] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3089c55d-f0d4-4da5-8dd2-cb69d6468528 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.188091] env[61006]: DEBUG nova.virt.block_device [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating existing volume attachment record: ccb702b0-5686-4e95-aade-f3524cf84c49 {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 944.351198] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.351917] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.354583] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.435s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.354823] env[61006]: DEBUG nova.objects.instance [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.365373] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337325, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572229} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.365653] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.365842] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.366236] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e4ec96e-e38b-45a1-9e99-bb686d2535b1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.373955] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 944.373955] env[61006]: value = "task-1337327" [ 944.373955] env[61006]: _type = "Task" [ 944.373955] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.384470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5711857a-5874-44f6-830c-640521164b99 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.074s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.385138] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.399124] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.399490] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance network_info: |[{"id": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "address": "fa:16:3e:b6:01:c1", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa62774d-28", "ovs_interfaceid": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 944.399806] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Acquired lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.400143] env[61006]: DEBUG nova.network.neutron [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Refreshing network info cache for port aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.401387] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:01:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa62774d-28ae-41d6-8bbb-3dae4be79437', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.409832] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating folder: Project (c4e87e7cb3f2444a8cc7bb7e65a99757). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.410448] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-672c2da6-06b8-4bc1-bd2c-3f17b22519c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.422515] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created folder: Project (c4e87e7cb3f2444a8cc7bb7e65a99757) in parent group-v285275. [ 944.422728] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating folder: Instances. Parent ref: group-v285396. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.422998] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a78df53b-a596-4efc-a070-82d7e9b4ac5c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.431153] env[61006]: DEBUG nova.compute.utils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.435837] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created folder: Instances in parent group-v285396. [ 944.436138] env[61006]: DEBUG oslo.service.loopingcall [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.436380] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.436632] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3112a45e-edb3-4a63-92d9-79c77c7b57e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.459224] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.459224] env[61006]: value = "task-1337330" [ 944.459224] env[61006]: _type = "Task" [ 944.459224] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.468487] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337330, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.530037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.858486] env[61006]: DEBUG nova.compute.utils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.860270] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 944.892368] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.135107} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.893029] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.894933] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c1da9f-c9b1-4f29-abf3-dc3c8dc51716 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.929475] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.933027] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb674a86-ec75-4fa6-9622-85d0b6913464 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.949579] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.022s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.959021] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 944.959021] env[61006]: value = "task-1337333" [ 944.959021] env[61006]: _type = "Task" [ 944.959021] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.973717] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337333, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.980864] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337330, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.030951] env[61006]: DEBUG nova.compute.manager [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.030951] env[61006]: DEBUG nova.compute.manager [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.030951] env[61006]: DEBUG oslo_concurrency.lockutils [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.031340] env[61006]: DEBUG oslo_concurrency.lockutils [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.032184] env[61006]: DEBUG nova.network.neutron [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.184938] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7877df-d2bc-4878-a9dc-0bc41bbcfac7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.194676] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82125ead-e101-4bf5-a4ae-13d17326823f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.230394] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0510a1-d91a-472c-95fd-a314b84eb449 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.238844] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fc084a-a5f0-4f06-9f18-6a33b1e3ee97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.255621] env[61006]: DEBUG nova.compute.provider_tree [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.358105] env[61006]: DEBUG nova.network.neutron [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updated VIF entry in instance network info cache for port aa62774d-28ae-41d6-8bbb-3dae4be79437. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.358499] env[61006]: DEBUG nova.network.neutron [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [{"id": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "address": "fa:16:3e:b6:01:c1", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa62774d-28", "ovs_interfaceid": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.364428] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.470658] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337333, 'name': ReconfigVM_Task, 'duration_secs': 0.345052} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.473944] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.474549] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337330, 'name': CreateVM_Task, 'duration_secs': 0.540068} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.474768] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-733bf520-56bf-432c-bfa7-694aa3215961 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.476248] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.476916] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.477091] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.477410] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.477991] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3e061c1-8066-4ed9-9914-8c3cf1cd9515 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.483645] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 945.483645] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526f3723-39c4-3b76-bba5-c6cb312801ec" [ 945.483645] env[61006]: _type = "Task" [ 945.483645] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.485109] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 945.485109] env[61006]: value = "task-1337334" [ 945.485109] env[61006]: _type = "Task" [ 945.485109] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.497443] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337334, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.502484] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526f3723-39c4-3b76-bba5-c6cb312801ec, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.504867] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.505205] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.507754] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.507754] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.507754] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.507754] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-418fd65e-71cd-4989-ba59-8fc99fc50774 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.516819] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.518512] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.518512] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-047799e0-7820-48aa-8ba1-ee872e523df8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.525011] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 945.525011] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522fbe0a-b083-2d0e-5b68-ca50b0e9df64" [ 945.525011] env[61006]: _type = "Task" [ 945.525011] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.535558] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522fbe0a-b083-2d0e-5b68-ca50b0e9df64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.758556] env[61006]: DEBUG nova.scheduler.client.report [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.783292] env[61006]: DEBUG nova.network.neutron [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.783636] env[61006]: DEBUG nova.network.neutron [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.861319] env[61006]: DEBUG oslo_concurrency.lockutils [req-b5ba02c6-4c6d-4949-b354-f744fb3f4f73 req-dcf4a4a9-d3e4-47a9-bf27-a75870a2fd03 service nova] Releasing lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.002665] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337334, 'name': Rename_Task, 'duration_secs': 0.160987} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.002746] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.003342] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6338f74-24e7-4303-83fb-eca15a99c996 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.007672] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.007926] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.008169] env[61006]: INFO nova.compute.manager [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Attaching volume d7571e01-7903-40f1-8588-89faa7a25ab9 to /dev/sdb [ 946.011994] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 946.011994] env[61006]: value = "task-1337335" [ 946.011994] env[61006]: _type = "Task" [ 946.011994] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.023761] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.036991] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522fbe0a-b083-2d0e-5b68-ca50b0e9df64, 'name': SearchDatastore_Task, 'duration_secs': 0.012027} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.038323] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71492c8a-2cc3-44e1-908e-eb9270faa79d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.047973] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 946.047973] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf6900-71eb-079a-339b-3003d36a9941" [ 946.047973] env[61006]: _type = "Task" [ 946.047973] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.065638] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf6900-71eb-079a-339b-3003d36a9941, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.067312] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a61fa0f-1e0a-4e84-891e-6b5ac4d877b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.076451] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65855294-b58b-4124-8dae-545344612a54 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.095349] env[61006]: DEBUG nova.virt.block_device [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating existing volume attachment record: dbd11e5a-9140-4ad3-be7f-f0a4332af30c {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 946.263372] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.265734] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.527s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.267438] env[61006]: INFO nova.compute.claims [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.294018] env[61006]: INFO nova.scheduler.client.report [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1 [ 946.297527] env[61006]: DEBUG oslo_concurrency.lockutils [req-691fba83-dc79-428e-9570-a4896a3283a2 req-0b312728-a939-468a-b8c1-c57482961689 service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.377163] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.403872] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.406478] env[61006]: DEBUG nova.virt.hardware [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.407451] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1a2bca-0982-4053-bfdc-7f2f067bef6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.418655] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987a964d-8c0a-4571-86bf-0a6bf6d0cd6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.434880] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.440469] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Creating folder: Project (7d0041ee4b13400e8b975ca94a871c0e). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.440811] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efbbfbe8-936b-4079-b4f2-39380d718e0d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.464658] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Created folder: Project (7d0041ee4b13400e8b975ca94a871c0e) in parent group-v285275. [ 946.464658] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Creating folder: Instances. Parent ref: group-v285402. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.464658] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-271d6c97-6c0b-4064-a014-ee13c0d01d80 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.473990] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Created folder: Instances in parent group-v285402. [ 946.474278] env[61006]: DEBUG oslo.service.loopingcall [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.474505] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.474738] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4c396ca-cc74-43b8-a83e-ea848b9d66b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.495583] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.495583] env[61006]: value = "task-1337340" [ 946.495583] env[61006]: _type = "Task" [ 946.495583] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.505902] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337340, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.524277] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337335, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.562337] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bf6900-71eb-079a-339b-3003d36a9941, 'name': SearchDatastore_Task, 'duration_secs': 0.013722} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.562644] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.562913] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 02408ca9-c580-444a-9608-a752146ca499/02408ca9-c580-444a-9608-a752146ca499.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.563272] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d05e1813-224f-4d6b-bdec-f60f0ee0da89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.566111] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.566391] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.575773] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 946.575773] env[61006]: value = "task-1337342" [ 946.575773] env[61006]: _type = "Task" [ 946.575773] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.592419] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.802619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-61a368f7-677a-49a7-9df2-9d1925ab5296 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.842s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.007510] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337340, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.026777] env[61006]: DEBUG oslo_vmware.api [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337335, 'name': PowerOnVM_Task, 'duration_secs': 0.613526} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.026940] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.027142] env[61006]: INFO nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Took 7.98 seconds to spawn the instance on the hypervisor. [ 947.027373] env[61006]: DEBUG nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.028347] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3308e1e8-3aea-4979-b0df-bc30b3e7c569 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.066097] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.066097] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing instance network info cache due to event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.066393] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.066393] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.066595] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.075724] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.076069] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 947.089874] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337342, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.488348] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebb38c4-eebf-4d65-a1cb-a3acca19d545 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.497606] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b4f4a1-c871-4028-b678-e274cb10cfa1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.509689] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337340, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.533981] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc5f58d-587e-4941-a1c3-03bf0ef7afe4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.547310] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bd2001-21e8-496c-aa14-717c0e3a0cf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.553460] env[61006]: INFO nova.compute.manager [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Took 15.20 seconds to build instance. [ 947.565446] env[61006]: DEBUG nova.compute.provider_tree [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.596396] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337342, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64024} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.597016] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 02408ca9-c580-444a-9608-a752146ca499/02408ca9-c580-444a-9608-a752146ca499.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.597295] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.597549] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7acf7dfa-f8b1-4a68-b829-4ee214566ecc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.608464] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 947.608464] env[61006]: value = "task-1337344" [ 947.608464] env[61006]: _type = "Task" [ 947.608464] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.620487] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.881883] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updated VIF entry in instance network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.881883] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.935039] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.935039] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.008635] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337340, 'name': CreateVM_Task, 'duration_secs': 1.440319} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.008851] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.009291] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.009447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.009776] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.010053] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e1d4b53-f5fd-4300-816d-0c05d22fc55d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.015438] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 948.015438] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd88d-73c0-4ed8-85bf-db84bf0a4fe6" [ 948.015438] env[61006]: _type = "Task" [ 948.015438] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.024129] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd88d-73c0-4ed8-85bf-db84bf0a4fe6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.055925] env[61006]: DEBUG oslo_concurrency.lockutils [None req-45a8fb8c-12b5-4c7f-ad43-83229acaa7a2 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.714s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.068704] env[61006]: DEBUG nova.scheduler.client.report [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.117901] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.161397} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.118187] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.118966] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c79c95d-db1f-4ec4-af5f-191f04f0654f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.141651] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 02408ca9-c580-444a-9608-a752146ca499/02408ca9-c580-444a-9608-a752146ca499.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.141969] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e77c8094-1d1a-41b9-8519-524dbc8f4141 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.163375] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 948.163375] env[61006]: value = "task-1337345" [ 948.163375] env[61006]: _type = "Task" [ 948.163375] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.172941] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.384360] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.384655] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.384924] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing instance network info cache due to event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.385266] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.385526] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.385803] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.437258] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.526680] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528fd88d-73c0-4ed8-85bf-db84bf0a4fe6, 'name': SearchDatastore_Task, 'duration_secs': 0.013705} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.527095] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.527293] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.527528] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.527689] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.527884] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.528168] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39febc40-0592-466a-8077-1001c6de4cfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.540993] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.540993] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.540993] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b699373-3b75-4ebc-a5a0-779b67ad9b93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.545524] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 948.545524] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52073abd-9d9e-3472-b55b-850c92ad91a9" [ 948.545524] env[61006]: _type = "Task" [ 948.545524] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.553801] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52073abd-9d9e-3472-b55b-850c92ad91a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.573955] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.574668] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.577689] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.919s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.577689] env[61006]: DEBUG nova.objects.instance [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lazy-loading 'resources' on Instance uuid d174a8ec-867e-4fea-b878-2a9af1476949 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.675116] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.700560] env[61006]: DEBUG nova.compute.manager [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Stashing vm_state: active {{(pid=61006) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 948.960181] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.056985] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52073abd-9d9e-3472-b55b-850c92ad91a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009861} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.060292] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-157e062a-4287-47ab-968b-c5c18d208b65 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.066739] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 949.066739] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efb05b-b447-7370-1ac8-3346e4595691" [ 949.066739] env[61006]: _type = "Task" [ 949.066739] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.076021] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efb05b-b447-7370-1ac8-3346e4595691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.080457] env[61006]: DEBUG nova.compute.utils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.081832] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.081969] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.133014] env[61006]: DEBUG nova.policy [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cab601d814fe4ac282bc1b9ea5f5ff9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8afbe151c21743bfb40dc12ba384db28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.135379] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updated VIF entry in instance network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.135718] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.176738] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337345, 'name': ReconfigVM_Task, 'duration_secs': 0.640375} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.177163] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 02408ca9-c580-444a-9608-a752146ca499/02408ca9-c580-444a-9608-a752146ca499.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.179899] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fd85736-ef45-43fe-b21f-120d42f233b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.188727] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 949.188727] env[61006]: value = "task-1337347" [ 949.188727] env[61006]: _type = "Task" [ 949.188727] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.203133] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337347, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.220109] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.362492] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaef7bce-8d08-48e6-a2fc-f85e16466a70 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.372520] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a890da89-bc82-4ea0-9867-0b9e961e7a10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.404815] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d95b529-39cf-40f1-b85f-cf335842096e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.414115] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0daa7899-2a8f-4704-aa1e-80ddc72050be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.432287] env[61006]: DEBUG nova.compute.provider_tree [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.446435] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 949.447356] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e906c1cc-c1d2-4653-b667-2a41b0233613 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.456467] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 949.456635] env[61006]: ERROR oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk due to incomplete transfer. [ 949.456866] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d4d7aa47-a29f-4cba-9615-3ca88f881bda {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.464927] env[61006]: DEBUG oslo_vmware.rw_handles [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5208fc4b-9efe-801b-2639-1030ea0053b7/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 949.465147] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Uploaded image 00666797-346f-4a13-84a1-8a64b7da3432 to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 949.467503] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 949.467762] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af7b5c22-d55f-4d28-a13f-ff924a1ce250 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.475144] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 949.475144] env[61006]: value = "task-1337348" [ 949.475144] env[61006]: _type = "Task" [ 949.475144] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.478899] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Successfully created port: 1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.487121] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337348, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.579023] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52efb05b-b447-7370-1ac8-3346e4595691, 'name': SearchDatastore_Task, 'duration_secs': 0.011446} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.582032] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.582032] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.582032] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7415fb9-18f9-4032-aa70-ea7a99af5c15 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.585711] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.593696] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 949.593696] env[61006]: value = "task-1337349" [ 949.593696] env[61006]: _type = "Task" [ 949.593696] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.605501] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.637962] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.640283] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.640283] env[61006]: DEBUG nova.compute.manager [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.640283] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.640283] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.640283] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.704019] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337347, 'name': Rename_Task, 'duration_secs': 0.184059} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.704019] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.704019] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4d3b780e-6027-4cc2-bd1c-69f53f0a8c5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.711870] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 949.711870] env[61006]: value = "task-1337350" [ 949.711870] env[61006]: _type = "Task" [ 949.711870] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.724220] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.937692] env[61006]: DEBUG nova.scheduler.client.report [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.987731] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337348, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.104923] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471316} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.105876] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.106047] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.106310] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c648a3f-983f-47e7-aef4-a7c142a24fe4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.114143] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 950.114143] env[61006]: value = "task-1337351" [ 950.114143] env[61006]: _type = "Task" [ 950.114143] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.123328] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.223480] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.315666] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.416299] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.416685] env[61006]: DEBUG nova.network.neutron [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.444364] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.867s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.446580] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.392s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.446828] env[61006]: DEBUG nova.objects.instance [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lazy-loading 'resources' on Instance uuid 40500a0d-33f3-492a-9573-ed10e45642c5 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.464028] env[61006]: INFO nova.scheduler.client.report [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Deleted allocations for instance d174a8ec-867e-4fea-b878-2a9af1476949 [ 950.487256] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337348, 'name': Destroy_Task, 'duration_secs': 0.595604} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.487537] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Destroyed the VM [ 950.487775] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 950.488048] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5560ed96-ca58-4909-ba2e-d6db69b3fb91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.495673] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 950.495673] env[61006]: value = "task-1337352" [ 950.495673] env[61006]: _type = "Task" [ 950.495673] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.505967] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337352, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.596965] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.621883] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.622209] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.622380] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.622563] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.622709] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.622858] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.623080] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.623288] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.623482] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.623590] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.623764] env[61006]: DEBUG nova.virt.hardware [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.624861] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dce3dc-8561-4210-8ac0-85430c5f909a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.638294] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dad2b77-2a63-43ec-9ff3-716785f92bf4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.642695] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06402} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.643772] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 950.643990] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285404', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'name': 'volume-d7571e01-7903-40f1-8588-89faa7a25ab9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a546f17-2fb8-4b99-9001-98cc6fe76837', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'serial': 'd7571e01-7903-40f1-8588-89faa7a25ab9'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 950.644633] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.645493] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91895c16-cc42-4f02-9901-782348777ed5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.649088] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df67924a-197c-4421-b417-4ccd7d006dde {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.678045] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.688646] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc654d38-2f72-4a09-809c-522a4f5b43f9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.703739] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6276b7-f665-4722-ab20-a8f7797c9c44 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.730622] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-d7571e01-7903-40f1-8588-89faa7a25ab9/volume-d7571e01-7903-40f1-8588-89faa7a25ab9.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.731038] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 950.731038] env[61006]: value = "task-1337353" [ 950.731038] env[61006]: _type = "Task" [ 950.731038] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.734379] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80ff7131-4876-4edc-b390-26ff0aba6226 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.758417] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337353, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.762754] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.763062] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 950.763062] env[61006]: value = "task-1337354" [ 950.763062] env[61006]: _type = "Task" [ 950.763062] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.773676] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337354, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.919437] env[61006]: DEBUG nova.compute.manager [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Received event network-vif-plugged-1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.920304] env[61006]: DEBUG oslo_concurrency.lockutils [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.920304] env[61006]: DEBUG oslo_concurrency.lockutils [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.920304] env[61006]: DEBUG oslo_concurrency.lockutils [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.920304] env[61006]: DEBUG nova.compute.manager [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] No waiting events found dispatching network-vif-plugged-1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.920578] env[61006]: WARNING nova.compute.manager [req-02c3763c-387c-4104-83b8-8f5dffe8e776 req-9b0768ee-36a2-46bf-8264-2ae1c207c577 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Received unexpected event network-vif-plugged-1a7b1ed1-7ded-4bd0-880f-96439cf27513 for instance with vm_state building and task_state spawning. [ 950.920921] env[61006]: DEBUG oslo_concurrency.lockutils [req-ad01b9e3-751e-4b41-ab8d-2dce22bccd97 req-0b382504-7387-4e5d-b604-66e337addadb service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.971275] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a8a3211c-043d-42f2-bc1b-670054df94d6 tempest-AttachVolumeTestJSON-668284865 tempest-AttachVolumeTestJSON-668284865-project-member] Lock "d174a8ec-867e-4fea-b878-2a9af1476949" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.104s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.007545] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337352, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.032277] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Successfully updated port: 1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.216787] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cbfd59-f537-4b45-b885-9ec4c61dff48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.228198] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebbfba1-5e0e-4eea-95ca-c4274a02e186 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.237680] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.272499] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0cf109-be19-46e1-9d54-58c35287efe4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.287433] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4146a62d-d6d5-461e-b450-c4bede70c0d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.292188] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337354, 'name': ReconfigVM_Task, 'duration_secs': 0.394153} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.292502] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337353, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.292927] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-d7571e01-7903-40f1-8588-89faa7a25ab9/volume-d7571e01-7903-40f1-8588-89faa7a25ab9.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.298101] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ee48787-bf12-41f5-af66-4604e1260875 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.319708] env[61006]: DEBUG nova.compute.provider_tree [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.328655] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 951.328655] env[61006]: value = "task-1337356" [ 951.328655] env[61006]: _type = "Task" [ 951.328655] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.339170] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.510304] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337352, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.535445] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.535445] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.535713] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.606172] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.606532] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.606532] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Forcefully refreshing network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 951.735036] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.774386] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337353, 'name': ReconfigVM_Task, 'duration_secs': 0.700158} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.774644] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Reconfigured VM instance instance-0000005f to attach disk [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.775384] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-034f4501-fa83-48a2-b82a-4738a7ad51d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.782456] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 951.782456] env[61006]: value = "task-1337357" [ 951.782456] env[61006]: _type = "Task" [ 951.782456] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.791247] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337357, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.823369] env[61006]: DEBUG nova.scheduler.client.report [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.839142] env[61006]: DEBUG oslo_vmware.api [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337356, 'name': ReconfigVM_Task, 'duration_secs': 0.190076} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.839431] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285404', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'name': 'volume-d7571e01-7903-40f1-8588-89faa7a25ab9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a546f17-2fb8-4b99-9001-98cc6fe76837', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'serial': 'd7571e01-7903-40f1-8588-89faa7a25ab9'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 952.007979] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337352, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.077221] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.229785] env[61006]: DEBUG nova.network.neutron [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.237861] env[61006]: DEBUG oslo_vmware.api [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337350, 'name': PowerOnVM_Task, 'duration_secs': 2.072004} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.238133] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.238382] env[61006]: INFO nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Took 10.71 seconds to spawn the instance on the hypervisor. [ 952.238587] env[61006]: DEBUG nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.239454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85294ae0-212d-4ed3-9c67-b0165d93a9b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.292678] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337357, 'name': Rename_Task, 'duration_secs': 0.396363} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.292951] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.293232] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-009ac706-7885-41d2-8cf0-e34c0ebe582c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.300268] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 952.300268] env[61006]: value = "task-1337358" [ 952.300268] env[61006]: _type = "Task" [ 952.300268] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.309880] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.328436] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.330999] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.801s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.353399] env[61006]: INFO nova.scheduler.client.report [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Deleted allocations for instance 40500a0d-33f3-492a-9573-ed10e45642c5 [ 952.508632] env[61006]: DEBUG oslo_vmware.api [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337352, 'name': RemoveSnapshot_Task, 'duration_secs': 1.88984} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.508924] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 952.509175] env[61006]: INFO nova.compute.manager [None req-583ce79e-b1ea-4a31-b839-399e5393ac1c tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 19.09 seconds to snapshot the instance on the hypervisor. [ 952.733027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.733391] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Instance network_info: |[{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.734028] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:1c:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37333dc2-982e-45e9-9dda-0c18417d7fa6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a7b1ed1-7ded-4bd0-880f-96439cf27513', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.741332] env[61006]: DEBUG oslo.service.loopingcall [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.741594] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.741948] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9b5dfaa-15b0-4669-b3ff-00afd0241cd8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.768470] env[61006]: INFO nova.compute.manager [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Took 18.18 seconds to build instance. [ 952.771336] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.771336] env[61006]: value = "task-1337359" [ 952.771336] env[61006]: _type = "Task" [ 952.771336] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.780897] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337359, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.813608] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337358, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.836272] env[61006]: INFO nova.compute.claims [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.864420] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3d7fb80-436c-4f52-8240-a979142ee295 tempest-ImagesTestJSON-1787015104 tempest-ImagesTestJSON-1787015104-project-member] Lock "40500a0d-33f3-492a-9573-ed10e45642c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.887s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.876538] env[61006]: DEBUG nova.objects.instance [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'flavor' on Instance uuid 1a546f17-2fb8-4b99-9001-98cc6fe76837 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.948734] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.956516] env[61006]: DEBUG nova.compute.manager [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Received event network-changed-1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.956731] env[61006]: DEBUG nova.compute.manager [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Refreshing instance network info cache due to event network-changed-1a7b1ed1-7ded-4bd0-880f-96439cf27513. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.956962] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] Acquiring lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.957119] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] Acquired lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.957291] env[61006]: DEBUG nova.network.neutron [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Refreshing network info cache for port 1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.270245] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b2d145ff-9d72-42af-bbb0-215a165fcbce tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.697s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.281015] env[61006]: DEBUG nova.compute.manager [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Received event network-changed-aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.281217] env[61006]: DEBUG nova.compute.manager [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Refreshing instance network info cache due to event network-changed-aa62774d-28ae-41d6-8bbb-3dae4be79437. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.281411] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] Acquiring lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.281552] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] Acquired lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.281714] env[61006]: DEBUG nova.network.neutron [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Refreshing network info cache for port aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.289147] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337359, 'name': CreateVM_Task, 'duration_secs': 0.394686} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.289271] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.290117] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.290117] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.290559] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.291532] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21d640e3-1aeb-4ef5-9dec-ab89d27b7960 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.298790] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 953.298790] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cd9a63-535b-08dc-a6e2-2eeda254ea4e" [ 953.298790] env[61006]: _type = "Task" [ 953.298790] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.311635] env[61006]: DEBUG oslo_vmware.api [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337358, 'name': PowerOnVM_Task, 'duration_secs': 0.758716} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.315250] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.315466] env[61006]: INFO nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Took 6.94 seconds to spawn the instance on the hypervisor. [ 953.316118] env[61006]: DEBUG nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.316118] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cd9a63-535b-08dc-a6e2-2eeda254ea4e, 'name': SearchDatastore_Task, 'duration_secs': 0.010805} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.316840] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ddeb81-621e-4383-83c6-2513367e7d1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.319211] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.319439] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.319667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.319814] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.319989] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.320654] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e32b3d3-0988-4269-bc5b-60819503e6d8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.330244] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.330430] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.331447] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fff897c8-c5ec-4bdd-9c90-3ab7edb1e44a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.337328] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 953.337328] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f84929-aeed-3d54-1d10-125ed7d585cf" [ 953.337328] env[61006]: _type = "Task" [ 953.337328] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.342777] env[61006]: INFO nova.compute.resource_tracker [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating resource usage from migration fa68ecfe-b11e-4599-a1c8-6fbe885ad3d3 [ 953.350888] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f84929-aeed-3d54-1d10-125ed7d585cf, 'name': SearchDatastore_Task, 'duration_secs': 0.009973} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.352227] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f98b4fe-ee36-437f-8c40-1d06ce9e8330 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.358724] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 953.358724] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294447d-c797-701f-3625-10f504631c0d" [ 953.358724] env[61006]: _type = "Task" [ 953.358724] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.369425] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294447d-c797-701f-3625-10f504631c0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.381439] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c034ab49-2457-4820-8884-c17c7b599a6e tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.373s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.451329] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.451535] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updated the network info_cache for instance {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 953.451705] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.451947] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.452695] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.452695] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.452695] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.452818] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.452893] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 953.453049] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 953.602237] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01462241-03ce-445c-9186-6b8d5b66693e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.607327] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.607613] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.607826] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.608015] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.608224] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.612357] env[61006]: INFO nova.compute.manager [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Terminating instance [ 953.614953] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce610de8-3642-4de6-a30a-65ef481cbcdf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.619025] env[61006]: DEBUG nova.compute.manager [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.619222] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.620287] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b4dd2d-85ca-446a-94f9-13923636a200 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.659706] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.660885] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358e7b0c-b5b8-4d86-921c-41cefd4ea0e4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.663700] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e15c09a8-59a3-4447-a545-932bc48b40d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.673195] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe1af46-a834-4fea-bfa8-83216ac35e80 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.677544] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 953.677544] env[61006]: value = "task-1337361" [ 953.677544] env[61006]: _type = "Task" [ 953.677544] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.693894] env[61006]: DEBUG nova.compute.provider_tree [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.703497] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.838510] env[61006]: INFO nova.compute.manager [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Took 17.66 seconds to build instance. [ 953.871513] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5294447d-c797-701f-3625-10f504631c0d, 'name': SearchDatastore_Task, 'duration_secs': 0.010277} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.871778] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.872048] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.872507] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d299017-7114-4cf8-ae2a-cdd2167d4b6a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.880429] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 953.880429] env[61006]: value = "task-1337362" [ 953.880429] env[61006]: _type = "Task" [ 953.880429] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.884522] env[61006]: DEBUG nova.network.neutron [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updated VIF entry in instance network info cache for port 1a7b1ed1-7ded-4bd0-880f-96439cf27513. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.884890] env[61006]: DEBUG nova.network.neutron [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.891896] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337362, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.956470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.065222] env[61006]: DEBUG nova.network.neutron [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updated VIF entry in instance network info cache for port aa62774d-28ae-41d6-8bbb-3dae4be79437. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.065634] env[61006]: DEBUG nova.network.neutron [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [{"id": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "address": "fa:16:3e:b6:01:c1", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa62774d-28", "ovs_interfaceid": "aa62774d-28ae-41d6-8bbb-3dae4be79437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.088382] env[61006]: INFO nova.compute.manager [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Rescuing [ 954.088648] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.088803] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.088973] env[61006]: DEBUG nova.network.neutron [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.189519] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337361, 'name': PowerOffVM_Task, 'duration_secs': 0.298415} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.190082] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.190411] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.190870] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6f41a50-8f1d-4183-bf2a-3f6239c8f095 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.198153] env[61006]: DEBUG nova.scheduler.client.report [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.278791] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.278956] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.279030] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Deleting the datastore file [datastore2] b27fc628-1ac8-4283-bf6e-bcff1cbfe149 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.279354] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4357ee47-e214-4c9a-8ee6-23506a024c4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.289892] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for the task: (returnval){ [ 954.289892] env[61006]: value = "task-1337364" [ 954.289892] env[61006]: _type = "Task" [ 954.289892] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.302110] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.341017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-41c309b2-67fa-4bd7-bb4e-1d5489d7a46d tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.171s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.389266] env[61006]: DEBUG oslo_concurrency.lockutils [req-bf3197cc-a2d2-485d-8798-0a7c54d9b3d2 req-a28b2361-a345-41ad-a6a3-60ee0a8b89b3 service nova] Releasing lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.391558] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337362, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49829} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.391805] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.392032] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.392343] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69c4d253-4343-4a7a-b969-5a3eb5fd8d9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.400572] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 954.400572] env[61006]: value = "task-1337365" [ 954.400572] env[61006]: _type = "Task" [ 954.400572] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.410779] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.568911] env[61006]: DEBUG oslo_concurrency.lockutils [req-0bf16831-0780-48e4-95e1-b1fa519578b5 req-bbe13de7-8b80-4f29-97c0-340f89496847 service nova] Releasing lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.706133] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.373s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.706133] env[61006]: INFO nova.compute.manager [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Migrating [ 954.718424] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.759s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.720795] env[61006]: INFO nova.compute.claims [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.805950] env[61006]: DEBUG oslo_vmware.api [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Task: {'id': task-1337364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210709} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.806443] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.806909] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.806909] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.806909] env[61006]: INFO nova.compute.manager [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 1.19 seconds to destroy the instance on the hypervisor. [ 954.807124] env[61006]: DEBUG oslo.service.loopingcall [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.807435] env[61006]: DEBUG nova.compute.manager [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 954.807516] env[61006]: DEBUG nova.network.neutron [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.872322] env[61006]: DEBUG nova.network.neutron [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.911562] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082244} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.911936] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.913086] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5201d017-9159-45c2-93c9-6a343c2d1b1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.938659] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.939880] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1dc327ba-3111-4174-b45f-a70e87fce706 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.962452] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 954.962452] env[61006]: value = "task-1337366" [ 954.962452] env[61006]: _type = "Task" [ 954.962452] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.971538] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337366, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.192314] env[61006]: INFO nova.compute.manager [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Rebuilding instance [ 955.226124] env[61006]: DEBUG nova.compute.manager [req-3d8fac2e-424f-4a42-93f1-8012767a4a92 req-ce796eb2-55bf-482b-8403-35c33a0e896b service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Received event network-vif-deleted-7ff4563e-2a36-4112-a195-52151bb79f12 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.226124] env[61006]: INFO nova.compute.manager [req-3d8fac2e-424f-4a42-93f1-8012767a4a92 req-ce796eb2-55bf-482b-8403-35c33a0e896b service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Neutron deleted interface 7ff4563e-2a36-4112-a195-52151bb79f12; detaching it from the instance and deleting it from the info cache [ 955.226124] env[61006]: DEBUG nova.network.neutron [req-3d8fac2e-424f-4a42-93f1-8012767a4a92 req-ce796eb2-55bf-482b-8403-35c33a0e896b service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.234044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.234044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.234044] env[61006]: DEBUG nova.network.neutron [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.266233] env[61006]: DEBUG nova.compute.manager [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.267161] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4daf1d8-d007-48c4-aa41-d143fc8b84ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.376876] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.474537] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.685512] env[61006]: DEBUG nova.network.neutron [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.731805] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3915cade-b73b-494d-97cc-c45ae8266b8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.752400] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb07ad71-4679-464f-be88-b52ea0b805d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.779920] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.792338] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68506c83-e4d4-4034-b636-c9578c6c51e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.794144] env[61006]: DEBUG nova.compute.manager [req-3d8fac2e-424f-4a42-93f1-8012767a4a92 req-ce796eb2-55bf-482b-8403-35c33a0e896b service nova] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Detach interface failed, port_id=7ff4563e-2a36-4112-a195-52151bb79f12, reason: Instance b27fc628-1ac8-4283-bf6e-bcff1cbfe149 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 955.801600] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 955.801600] env[61006]: value = "task-1337367" [ 955.801600] env[61006]: _type = "Task" [ 955.801600] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.811483] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337367, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.906876] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.909037] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c4fd30d-bafa-4e6a-ad0b-6577a9359a48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.920249] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 955.920249] env[61006]: value = "task-1337368" [ 955.920249] env[61006]: _type = "Task" [ 955.920249] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.929949] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337368, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.982737] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337366, 'name': ReconfigVM_Task, 'duration_secs': 0.754999} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.983916] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfigured VM instance instance-00000060 to attach disk [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.983916] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ee62a7c-d5d5-479e-972c-b6baa634393b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.992648] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 955.992648] env[61006]: value = "task-1337369" [ 955.992648] env[61006]: _type = "Task" [ 955.992648] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.009800] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337369, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.031611] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c143735d-5d16-44c2-bbde-079f81508c45 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.040960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4275c3d6-e8a9-4a52-844f-fb4e0a93b0af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.081232] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24afadcc-a35f-4944-8c34-1d8f6478543b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.090620] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1780178c-c429-47fd-a6f2-44879b6b0a44 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.107295] env[61006]: DEBUG nova.compute.provider_tree [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.188761] env[61006]: INFO nova.compute.manager [-] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Took 1.38 seconds to deallocate network for instance. [ 956.292529] env[61006]: DEBUG nova.network.neutron [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.312619] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337367, 'name': PowerOffVM_Task, 'duration_secs': 0.215264} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.313640] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.313849] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.314653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5759db78-e94f-453e-ab9c-e8f907cae8e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.323112] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.323736] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22c74a00-5c22-4d90-a2d5-1eb9cb07e782 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.354677] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.354955] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.355182] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Deleting the datastore file [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.355778] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55ac047c-452f-45c0-8555-7a0a83a155c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.364647] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 956.364647] env[61006]: value = "task-1337371" [ 956.364647] env[61006]: _type = "Task" [ 956.364647] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.374018] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.430517] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337368, 'name': PowerOffVM_Task, 'duration_secs': 0.225184} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.430803] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.431607] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b441bb-3236-4238-b091-77c968f30c74 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.458408] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b6e5a7-c371-4482-922f-cad13f593fdc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.491107] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.491107] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-003d88e2-1dad-451e-9bb3-768ec751764e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.498458] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 956.498458] env[61006]: value = "task-1337372" [ 956.498458] env[61006]: _type = "Task" [ 956.498458] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.504317] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337369, 'name': Rename_Task, 'duration_secs': 0.206979} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.504900] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.505128] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54d95b24-bd4c-45e8-beb9-5c3438cb9225 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.510721] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 956.510913] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.511170] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.511331] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.511526] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.511809] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 956.511809] env[61006]: value = "task-1337373" [ 956.511809] env[61006]: _type = "Task" [ 956.511809] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.511993] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2ef2060-1d25-4e95-bc49-2a21fcf1f835 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.521409] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.523923] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.523923] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 956.524723] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0bbddcb-4cd5-4fae-8132-a5e84defeea1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.529766] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 956.529766] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f7d8b5-b5a7-63ff-fe17-c65f24c55dba" [ 956.529766] env[61006]: _type = "Task" [ 956.529766] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.538491] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f7d8b5-b5a7-63ff-fe17-c65f24c55dba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.611115] env[61006]: DEBUG nova.scheduler.client.report [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.695099] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.795862] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.875294] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100567} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.875894] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.876099] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.876283] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.026277] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337373, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.043626] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52f7d8b5-b5a7-63ff-fe17-c65f24c55dba, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.044549] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35743ce6-7783-4b63-a12a-bb030b06aec0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.052294] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 957.052294] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522a4dbc-de48-bfb2-785d-0631a6d90e3a" [ 957.052294] env[61006]: _type = "Task" [ 957.052294] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.062253] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522a4dbc-de48-bfb2-785d-0631a6d90e3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.118425] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.118425] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.121280] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.901s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.526737] env[61006]: DEBUG oslo_vmware.api [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337373, 'name': PowerOnVM_Task, 'duration_secs': 0.854672} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.527146] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.527249] env[61006]: INFO nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Took 6.93 seconds to spawn the instance on the hypervisor. [ 957.527405] env[61006]: DEBUG nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.528599] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3790cba-ca16-43ac-8860-2916b2595955 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.564884] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522a4dbc-de48-bfb2-785d-0631a6d90e3a, 'name': SearchDatastore_Task, 'duration_secs': 0.01732} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.568838] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.568838] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. {{(pid=61006) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 957.568838] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f44938ed-a09b-4223-b378-137d17922856 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.578517] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 957.578517] env[61006]: value = "task-1337374" [ 957.578517] env[61006]: _type = "Task" [ 957.578517] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.589666] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337374, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.625280] env[61006]: DEBUG nova.compute.utils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.626701] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.626964] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 957.632951] env[61006]: INFO nova.compute.claims [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.703831] env[61006]: DEBUG nova.policy [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.925802] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.926160] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.926252] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.926420] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.926566] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.926770] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.926918] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.929385] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.929696] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.929890] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.930099] env[61006]: DEBUG nova.virt.hardware [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.931063] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3bacf6-2060-46a3-bd8a-4065cfca09a1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.942402] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e200d4e-c998-4127-a696-2c576f7f54f9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.959844] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.966921] env[61006]: DEBUG oslo.service.loopingcall [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.967338] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.967666] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed2b0659-7010-4b54-af45-bc4c5637b6e2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.990948] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.990948] env[61006]: value = "task-1337375" [ 957.990948] env[61006]: _type = "Task" [ 957.990948] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.001327] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337375, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.053295] env[61006]: INFO nova.compute.manager [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Took 20.33 seconds to build instance. [ 958.061944] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Successfully created port: bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.094425] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337374, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512316} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.094890] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk. [ 958.096037] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbebe98c-6cf0-4e2b-b876-91fcde2973cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.128879] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 958.129280] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dacbffd6-1896-4ca3-971f-573cf67562bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.142677] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.149434] env[61006]: INFO nova.compute.resource_tracker [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating resource usage from migration 8e633730-59d9-4a02-be82-40f008be263d [ 958.157762] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 958.157762] env[61006]: value = "task-1337376" [ 958.157762] env[61006]: _type = "Task" [ 958.157762] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.169775] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337376, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.316298] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbc41e2-bf0c-424f-9258-530995cf30f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.341013] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 0 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.444531] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f938e9-e8f2-4147-9691-f6d6db68ee02 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.453679] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1728a6-2780-401e-9919-d1bbbd3555f1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.487283] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0daf8145-5555-43c0-9cf6-a83d087d4b3f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.499663] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c016504-bca5-48a0-915b-87133b82ea42 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.512154] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337375, 'name': CreateVM_Task, 'duration_secs': 0.381529} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.512154] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.512154] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.512154] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.512154] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.512154] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03277ee8-6e56-4e96-ba83-8a42d0fd585b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.521885] env[61006]: DEBUG nova.compute.provider_tree [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.528481] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 958.528481] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e227a6-07fd-b553-be0f-b8d897e4a16b" [ 958.528481] env[61006]: _type = "Task" [ 958.528481] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.538953] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e227a6-07fd-b553-be0f-b8d897e4a16b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.555430] env[61006]: DEBUG oslo_concurrency.lockutils [None req-be3c08cd-1e0f-43ac-9f4d-17a00816ccef tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.842s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.668916] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337376, 'name': ReconfigVM_Task, 'duration_secs': 0.428184} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.669091] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0-rescue.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.669940] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2f65b1-28a4-4253-bcfc-beacf99d9cfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.699040] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3020c26d-ee98-45cc-a41c-e9e119979c19 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.716165] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 958.716165] env[61006]: value = "task-1337377" [ 958.716165] env[61006]: _type = "Task" [ 958.716165] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.725541] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.849622] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.849959] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db86262d-4b0b-45b5-9f90-6db5ee6a14b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.861070] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 958.861070] env[61006]: value = "task-1337378" [ 958.861070] env[61006]: _type = "Task" [ 958.861070] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.875124] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 958.875324] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 17 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 959.025452] env[61006]: DEBUG nova.scheduler.client.report [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.041088] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e227a6-07fd-b553-be0f-b8d897e4a16b, 'name': SearchDatastore_Task, 'duration_secs': 0.024371} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.041418] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.041647] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.041885] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.042697] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.043241] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.044034] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.044294] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.044730] env[61006]: DEBUG nova.objects.instance [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'flavor' on Instance uuid bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.046697] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ce99f78-4054-4755-940e-3e49eb8f284c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.058174] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.058174] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.058481] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f86570c8-13de-4016-9938-41dbf16a5d54 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.067362] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 959.067362] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98227-86a8-0f61-1882-7a4bbec9640d" [ 959.067362] env[61006]: _type = "Task" [ 959.067362] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.078860] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98227-86a8-0f61-1882-7a4bbec9640d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.158468] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.210373] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.210707] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.211505] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.211505] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.211505] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.211731] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.212042] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.212303] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.212600] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.212821] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.213199] env[61006]: DEBUG nova.virt.hardware [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.214236] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2720be1-f1f2-4780-a4e7-48d765312882 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.230071] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337377, 'name': ReconfigVM_Task, 'duration_secs': 0.18697} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.230071] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.231224] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75287156-ebef-4187-b692-41fbfadb6da3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.232954] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33d6f2a-a68b-4634-b546-b9a25b20c504 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.249524] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 959.249524] env[61006]: value = "task-1337379" [ 959.249524] env[61006]: _type = "Task" [ 959.249524] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.266401] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.389154] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.389318] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.389497] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.389704] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.389854] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.390219] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.394331] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.394544] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.394718] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.394888] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.395087] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.400434] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c683cf1e-4ec3-44d7-a621-2c18b3c0f711 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.418647] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 959.418647] env[61006]: value = "task-1337380" [ 959.418647] env[61006]: _type = "Task" [ 959.418647] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.430086] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.533947] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.413s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.534359] env[61006]: INFO nova.compute.manager [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Migrating [ 959.543967] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.228s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.544221] env[61006]: DEBUG nova.objects.instance [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'pci_requests' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.573252] env[61006]: DEBUG nova.objects.instance [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'numa_topology' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.593120] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b98227-86a8-0f61-1882-7a4bbec9640d, 'name': SearchDatastore_Task, 'duration_secs': 0.011464} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.594221] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3edf90ad-aa60-425e-88e5-7710671fd8c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.601987] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 959.601987] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270b9c1-25ae-8ab0-072d-f21d8a4abe98" [ 959.601987] env[61006]: _type = "Task" [ 959.601987] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.611606] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270b9c1-25ae-8ab0-072d-f21d8a4abe98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.745283] env[61006]: DEBUG nova.objects.instance [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'pci_requests' on Instance uuid bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.764560] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337379, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.843126] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Successfully updated port: bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.930464] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337380, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.004908] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.005340] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.076407] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.076715] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.077438] env[61006]: DEBUG nova.network.neutron [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.078402] env[61006]: INFO nova.compute.claims [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.115300] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5270b9c1-25ae-8ab0-072d-f21d8a4abe98, 'name': SearchDatastore_Task, 'duration_secs': 0.010683} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.115772] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.116391] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.116662] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b2e72a8-6a87-45cd-ba75-8f11b369460c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.127616] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 960.127616] env[61006]: value = "task-1337381" [ 960.127616] env[61006]: _type = "Task" [ 960.127616] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.138120] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.247660] env[61006]: DEBUG nova.objects.base [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 960.248479] env[61006]: DEBUG nova.network.neutron [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.263817] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337379, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.349235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.349370] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.349502] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.352779] env[61006]: DEBUG nova.policy [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.361222] env[61006]: DEBUG nova.compute.manager [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Received event network-vif-plugged-bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.361514] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] Acquiring lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.361693] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.361954] env[61006]: DEBUG oslo_concurrency.lockutils [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.362272] env[61006]: DEBUG nova.compute.manager [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] No waiting events found dispatching network-vif-plugged-bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.362439] env[61006]: WARNING nova.compute.manager [req-ce70c76a-efb8-4b60-9433-0d05b32fa683 req-5e03219b-1ab0-488b-98a6-bb67637060cd service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Received unexpected event network-vif-plugged-bfba9043-9516-43cb-9874-048f6b5076b7 for instance with vm_state building and task_state spawning. [ 960.432998] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337380, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.509283] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.645078] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337381, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.725775] env[61006]: DEBUG nova.compute.manager [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Stashing vm_state: active {{(pid=61006) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 960.767802] env[61006]: DEBUG oslo_vmware.api [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337379, 'name': PowerOnVM_Task, 'duration_secs': 1.039839} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.771809] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.776381] env[61006]: DEBUG nova.compute.manager [None req-7e5da110-942e-459b-ad41-30d66e919230 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.777324] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b98449-56af-4926-b75d-3258bff2a96c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.859119] env[61006]: DEBUG nova.network.neutron [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.883919] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 960.933050] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337380, 'name': ReconfigVM_Task, 'duration_secs': 1.231529} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.933386] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 33 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.029433] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.038267] env[61006]: DEBUG nova.network.neutron [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Updating instance_info_cache with network_info: [{"id": "bfba9043-9516-43cb-9874-048f6b5076b7", "address": "fa:16:3e:f9:47:64", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfba9043-95", "ovs_interfaceid": "bfba9043-9516-43cb-9874-048f6b5076b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.138546] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.782583} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.138814] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.139028] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.139324] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfae293c-4cb7-448a-8cc9-520e7f905cea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.148442] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 961.148442] env[61006]: value = "task-1337382" [ 961.148442] env[61006]: _type = "Task" [ 961.148442] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.160268] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.241084] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.321912] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37addc82-9808-443c-bd8c-4044b0c06d53 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.330785] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381e4094-4a34-45a7-8bfa-47b236bac8eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.363143] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.365184] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d19599-b93e-4f7c-b2d6-3f0881df4cdb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.374160] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcf4041-e224-4cf4-bd65-fd00ac61297d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.389546] env[61006]: DEBUG nova.compute.provider_tree [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.440169] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.440405] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.441194] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.441194] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.441194] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.441194] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.441417] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.441459] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.441602] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.441850] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.442117] env[61006]: DEBUG nova.virt.hardware [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.447591] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 961.447891] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e25360d1-fe70-487a-9ad8-6df2746b9b48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.469181] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 961.469181] env[61006]: value = "task-1337383" [ 961.469181] env[61006]: _type = "Task" [ 961.469181] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.478214] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.541624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.541624] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Instance network_info: |[{"id": "bfba9043-9516-43cb-9874-048f6b5076b7", "address": "fa:16:3e:f9:47:64", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfba9043-95", "ovs_interfaceid": "bfba9043-9516-43cb-9874-048f6b5076b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.541914] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:47:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfba9043-9516-43cb-9874-048f6b5076b7', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.550019] env[61006]: DEBUG oslo.service.loopingcall [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.550395] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.550761] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c9bed44-e958-456a-b48f-e20a87386ee9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.579221] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.579221] env[61006]: value = "task-1337384" [ 961.579221] env[61006]: _type = "Task" [ 961.579221] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.589318] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337384, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.661157] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07246} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.661157] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.662068] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd8ea62-b298-434a-9487-3eefb2e93b62 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.684388] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.685037] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bcd587f-3de4-4134-a25e-997ec7f27b39 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.707551] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 961.707551] env[61006]: value = "task-1337385" [ 961.707551] env[61006]: _type = "Task" [ 961.707551] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.716232] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337385, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.897020] env[61006]: DEBUG nova.scheduler.client.report [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.982022] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337383, 'name': ReconfigVM_Task, 'duration_secs': 0.402994} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.982022] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.982022] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbeb8891-77cb-4187-a690-9c55c3838208 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.006616] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.007278] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03de822b-8984-4113-9daa-b53968b48904 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.028638] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 962.028638] env[61006]: value = "task-1337386" [ 962.028638] env[61006]: _type = "Task" [ 962.028638] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.038101] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337386, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.089937] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337384, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.105006] env[61006]: DEBUG nova.network.neutron [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Successfully updated port: f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.143996] env[61006]: INFO nova.compute.manager [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Unrescuing [ 962.144321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.144575] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.144777] env[61006]: DEBUG nova.network.neutron [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.219954] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337385, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.399431] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.855s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.401663] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.445s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.401831] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.401982] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 962.404044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.707s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.404044] env[61006]: DEBUG nova.objects.instance [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lazy-loading 'resources' on Instance uuid b27fc628-1ac8-4283-bf6e-bcff1cbfe149 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.405218] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ff2491-f9bc-4536-8d74-113fe84c01f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.415884] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86684ef1-d488-4736-a838-cb03e8684388 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.433168] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667b502b-1275-48b7-86ba-6476ad96515a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.440431] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d905dbbe-e087-45c2-8b3e-a82f1d2daa40 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.488711] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179742MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 962.488916] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.539945] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.591421] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337384, 'name': CreateVM_Task, 'duration_secs': 0.844816} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.591822] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.593037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.593393] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.593860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.594338] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0834b4d7-92ca-40f4-a42a-a7441dcdbaae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.601026] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 962.601026] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522f8d3c-6b2f-6123-1446-237150115c30" [ 962.601026] env[61006]: _type = "Task" [ 962.601026] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.612974] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.613338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.613592] env[61006]: DEBUG nova.network.neutron [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.615059] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522f8d3c-6b2f-6123-1446-237150115c30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.667437] env[61006]: INFO nova.network.neutron [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating port 9230a732-200c-4084-8c6c-a5892e9a50ba with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 962.698799] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Received event network-changed-bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.699014] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Refreshing instance network info cache due to event network-changed-bfba9043-9516-43cb-9874-048f6b5076b7. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.699236] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Acquiring lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.699382] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Acquired lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.699542] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Refreshing network info cache for port bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.719454] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337385, 'name': ReconfigVM_Task, 'duration_secs': 0.927235} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.722285] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Reconfigured VM instance instance-0000005f to attach disk [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c/a1458c3b-f6d4-4cde-ad9a-4e4693ff312c.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.722910] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-018352a0-3bed-4230-9f6b-56c9fead5379 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.731297] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 962.731297] env[61006]: value = "task-1337387" [ 962.731297] env[61006]: _type = "Task" [ 962.731297] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.739527] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337387, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.884792] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb9b9d4-c402-43ae-b06c-3f3257e60f59 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.889243] env[61006]: DEBUG nova.network.neutron [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.911412] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 0 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.044217] env[61006]: DEBUG oslo_vmware.api [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337386, 'name': ReconfigVM_Task, 'duration_secs': 0.68474} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.044364] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01/524f3fd1-1e71-40c0-96c2-0acac5055e01.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.044609] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 50 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.113833] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522f8d3c-6b2f-6123-1446-237150115c30, 'name': SearchDatastore_Task, 'duration_secs': 0.01096} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.114164] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.114692] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.114692] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.114809] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.114938] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.115440] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdff3f50-f259-4ef3-b835-871eba93c55d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.133592] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.133969] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.134574] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aad9fce-9e72-49a5-96ce-2bed9ef72394 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.140974] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 963.140974] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5272a99a-cb93-d299-cac3-7b4bb191792a" [ 963.140974] env[61006]: _type = "Task" [ 963.140974] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.150221] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5272a99a-cb93-d299-cac3-7b4bb191792a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.151349] env[61006]: WARNING nova.network.neutron [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] 98dcba29-9f32-40eb-b526-46e091c2a330 already exists in list: networks containing: ['98dcba29-9f32-40eb-b526-46e091c2a330']. ignoring it [ 963.169044] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329f3400-fabe-4fd0-8dd6-4d73927b0824 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.177831] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e94989-1c08-46c8-9c8a-6fa9f98a9b3c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.210621] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd6e442-0c6a-41b0-8df9-d0dfd623d889 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.219255] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbb0123-d326-4dd3-908b-b5cec5069ab3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.233904] env[61006]: DEBUG nova.compute.provider_tree [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.242984] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337387, 'name': Rename_Task, 'duration_secs': 0.246215} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.243905] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.244167] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b37e7822-fd18-453c-acdb-97ded72a3056 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.256129] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 963.256129] env[61006]: value = "task-1337388" [ 963.256129] env[61006]: _type = "Task" [ 963.256129] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.265916] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337388, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.393053] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.393781] env[61006]: DEBUG nova.objects.instance [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'flavor' on Instance uuid 1a546f17-2fb8-4b99-9001-98cc6fe76837 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.417608] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.418127] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecb73707-2859-41af-899d-3e4c80215fee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.427096] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 963.427096] env[61006]: value = "task-1337389" [ 963.427096] env[61006]: _type = "Task" [ 963.427096] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.436772] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.549851] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Updated VIF entry in instance network info cache for port bfba9043-9516-43cb-9874-048f6b5076b7. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.550252] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Updating instance_info_cache with network_info: [{"id": "bfba9043-9516-43cb-9874-048f6b5076b7", "address": "fa:16:3e:f9:47:64", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfba9043-95", "ovs_interfaceid": "bfba9043-9516-43cb-9874-048f6b5076b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.553642] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44b7816-033c-46a1-aacb-072b026c192e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.578226] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb33dcc-4062-4bc2-86b4-22f0dde8f730 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.600483] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 67 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.654896] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5272a99a-cb93-d299-cac3-7b4bb191792a, 'name': SearchDatastore_Task, 'duration_secs': 0.035442} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.655716] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-355d988f-77ea-44b1-a817-7303c70f8c8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.663262] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 963.663262] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f8cdb-cc45-62ac-0f12-b63acc1ae34a" [ 963.663262] env[61006]: _type = "Task" [ 963.663262] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.671934] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f8cdb-cc45-62ac-0f12-b63acc1ae34a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.739150] env[61006]: DEBUG nova.scheduler.client.report [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.767009] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337388, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.822823] env[61006]: DEBUG nova.network.neutron [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "address": "fa:16:3e:0a:c8:a3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07eb7f2-10", "ovs_interfaceid": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.899630] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b232562f-ec24-44de-87a1-3b7ac39ae22d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.926166] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.926544] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06948b06-1b14-48db-938f-baabcf971759 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.937113] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337389, 'name': PowerOffVM_Task, 'duration_secs': 0.244376} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.938759] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.938759] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 17 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.942345] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 963.942345] env[61006]: value = "task-1337390" [ 963.942345] env[61006]: _type = "Task" [ 963.942345] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.951850] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.053785] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Releasing lock "refresh_cache-0b2b6d52-8340-4f85-9fc0-ab1f2b079031" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.054152] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.054402] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.054820] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.055131] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.055459] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] No waiting events found dispatching network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.055606] env[61006]: WARNING nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received unexpected event network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 for instance with vm_state active and task_state None. [ 964.055817] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.056030] env[61006]: DEBUG nova.compute.manager [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-f07eb7f2-10e5-4957-a5c8-3792f6d15b26. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.056347] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.148516] env[61006]: DEBUG nova.network.neutron [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Port 3d9ba3b1-a27e-4002-bd16-cfd881fca452 binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 964.175039] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]524f8cdb-cc45-62ac-0f12-b63acc1ae34a, 'name': SearchDatastore_Task, 'duration_secs': 0.01097} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.175426] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.175583] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 0b2b6d52-8340-4f85-9fc0-ab1f2b079031/0b2b6d52-8340-4f85-9fc0-ab1f2b079031.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.175851] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e3868d7-73fa-4a5e-937b-67f95060ab4e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.184231] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 964.184231] env[61006]: value = "task-1337391" [ 964.184231] env[61006]: _type = "Task" [ 964.184231] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.193629] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.243791] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.841s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.246236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.217s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.247757] env[61006]: INFO nova.compute.claims [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.266937] env[61006]: INFO nova.scheduler.client.report [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Deleted allocations for instance b27fc628-1ac8-4283-bf6e-bcff1cbfe149 [ 964.276786] env[61006]: DEBUG oslo_vmware.api [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337388, 'name': PowerOnVM_Task, 'duration_secs': 0.778859} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.277257] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.277467] env[61006]: DEBUG nova.compute.manager [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.278300] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a3485a-1219-4b5d-afc1-5c937a6c7e52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.325376] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.326022] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.326188] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.326692] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.326884] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.328836] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b82d460-9f60-4a2f-980f-5a20b65a91ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.347351] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.347609] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.347761] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.347945] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.348101] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.348255] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.348467] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.348630] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.348800] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.348962] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.349152] env[61006]: DEBUG nova.virt.hardware [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.355852] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfiguring VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 964.356553] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cb5b117-d995-483a-bf8b-3cabccf48bbb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.376463] env[61006]: DEBUG oslo_vmware.api [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 964.376463] env[61006]: value = "task-1337392" [ 964.376463] env[61006]: _type = "Task" [ 964.376463] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.387434] env[61006]: DEBUG oslo_vmware.api [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337392, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.397805] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.398070] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.398316] env[61006]: DEBUG nova.network.neutron [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.445787] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.445887] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.446032] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.446225] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.446384] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.446568] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.446782] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.446944] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.447137] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.447304] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.447482] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.453068] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87950737-51b1-4f54-a4ec-968b3be0c8e2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.476417] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337390, 'name': PowerOffVM_Task, 'duration_secs': 0.250802} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.477971] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.483378] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfiguring VM instance instance-00000055 to detach disk 2002 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 964.483818] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 964.483818] env[61006]: value = "task-1337393" [ 964.483818] env[61006]: _type = "Task" [ 964.483818] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.484060] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d68f813e-8104-424e-8468-595dc0fb5490 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.512019] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.513693] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 964.513693] env[61006]: value = "task-1337394" [ 964.513693] env[61006]: _type = "Task" [ 964.513693] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.525634] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337394, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.694602] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337391, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.769229] env[61006]: DEBUG nova.compute.manager [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.769487] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.769702] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.769858] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.769983] env[61006]: DEBUG nova.compute.manager [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] No waiting events found dispatching network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.770165] env[61006]: WARNING nova.compute.manager [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received unexpected event network-vif-plugged-9230a732-200c-4084-8c6c-a5892e9a50ba for instance with vm_state shelved_offloaded and task_state spawning. [ 964.770380] env[61006]: DEBUG nova.compute.manager [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.770501] env[61006]: DEBUG nova.compute.manager [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing instance network info cache due to event network-changed-9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.770667] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Acquiring lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.780360] env[61006]: DEBUG oslo_concurrency.lockutils [None req-388e2730-559d-4e39-9fa8-47306f3d060a tempest-ImagesOneServerTestJSON-911145136 tempest-ImagesOneServerTestJSON-911145136-project-member] Lock "b27fc628-1ac8-4283-bf6e-bcff1cbfe149" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.173s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.794044] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.885994] env[61006]: DEBUG oslo_vmware.api [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337392, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.008943] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337393, 'name': ReconfigVM_Task, 'duration_secs': 0.183642} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.011141] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 33 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.032018] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337394, 'name': ReconfigVM_Task, 'duration_secs': 0.365647} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.032670] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfigured VM instance instance-00000055 to detach disk 2002 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 965.032974] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.033632] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4458053c-b211-4ac0-9c02-5397899d4233 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.045642] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 965.045642] env[61006]: value = "task-1337395" [ 965.045642] env[61006]: _type = "Task" [ 965.045642] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.056618] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.141069] env[61006]: DEBUG nova.network.neutron [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.172514] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.172774] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.172970] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.199806] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514792} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.200208] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 0b2b6d52-8340-4f85-9fc0-ab1f2b079031/0b2b6d52-8340-4f85-9fc0-ab1f2b079031.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 965.200350] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.200554] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0c96147-84b5-45ff-a366-9f4d756ff92e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.209418] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 965.209418] env[61006]: value = "task-1337396" [ 965.209418] env[61006]: _type = "Task" [ 965.209418] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.223156] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.245853] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.246186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.246445] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.246742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.246952] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.249259] env[61006]: INFO nova.compute.manager [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Terminating instance [ 965.251413] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "refresh_cache-a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.251644] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquired lock "refresh_cache-a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.251834] env[61006]: DEBUG nova.network.neutron [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.264105] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port f07eb7f2-10e5-4957-a5c8-3792f6d15b26. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.264769] env[61006]: DEBUG nova.network.neutron [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "address": "fa:16:3e:0a:c8:a3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07eb7f2-10", "ovs_interfaceid": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.388285] env[61006]: DEBUG oslo_vmware.api [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337392, 'name': ReconfigVM_Task, 'duration_secs': 0.921888} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.391440] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.391687] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfigured VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 965.520455] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.520709] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.520871] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.521069] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.521224] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.521377] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.521634] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.521742] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.521915] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.522085] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.522274] env[61006]: DEBUG nova.virt.hardware [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.527641] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 965.528896] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-218b5cc6-acc2-4d6c-a24d-2c6d31375813 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.542329] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53f8035-c79f-4d09-81ea-17980ced4e97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.554242] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce747f3a-b6e6-4677-bbe0-564106770b81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.557480] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 965.557480] env[61006]: value = "task-1337397" [ 965.557480] env[61006]: _type = "Task" [ 965.557480] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.588694] env[61006]: DEBUG oslo_vmware.api [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337395, 'name': PowerOnVM_Task, 'duration_secs': 0.45872} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.589695] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 965.589896] env[61006]: DEBUG nova.compute.manager [None req-6e60b11f-72b8-4803-9954-aae8993a67f4 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.590695] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659c8058-d6af-4563-9640-45c9042338ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.596802] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c09a8e-f354-47bc-b6eb-a4bd35e04e05 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.599475] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337397, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.609526] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdf2243-80ef-43c9-b276-594046d50127 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.626014] env[61006]: DEBUG nova.compute.provider_tree [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.644327] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.647164] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Acquired lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.647461] env[61006]: DEBUG nova.network.neutron [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Refreshing network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.675737] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8bfec3a945f7135114623b3fc0cc2bb0',container_format='bare',created_at=2024-10-22T16:15:17Z,direct_url=,disk_format='vmdk',id=87693d49-a593-43f4-8a3e-5dc9080f7010,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-608688077-shelved',owner='f60c05599bb7457f9bd7a3d11daf9ab3',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-22T16:15:32Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.675993] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.676171] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.676362] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.676510] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.676761] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.676975] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.677520] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.677520] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.677520] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.677722] env[61006]: DEBUG nova.virt.hardware [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.682909] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dfb127-bc11-403e-af6d-320172426806 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.690314] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934c9d93-0a4c-4b6e-8c0c-6e32258c3d12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.705455] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:76:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9230a732-200c-4084-8c6c-a5892e9a50ba', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.713177] env[61006]: DEBUG oslo.service.loopingcall [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.713938] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.717707] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c74cf23-8d8f-4038-8c77-d719c63fe8aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.743666] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113424} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.745212] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.745346] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.745346] env[61006]: value = "task-1337398" [ 965.745346] env[61006]: _type = "Task" [ 965.745346] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.748414] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae2fba8-5919-46c6-b323-03d451d5f366 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.758369] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337398, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.768291] env[61006]: DEBUG oslo_concurrency.lockutils [req-5fa61c84-50f7-4382-a61e-427114e2efd9 req-b8fb0a3b-7e58-4bfb-bde1-c7718267ce51 service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.779126] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 0b2b6d52-8340-4f85-9fc0-ab1f2b079031/0b2b6d52-8340-4f85-9fc0-ab1f2b079031.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.779126] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0add5126-ef28-41ee-b637-bc21e2746474 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.798853] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 965.798853] env[61006]: value = "task-1337399" [ 965.798853] env[61006]: _type = "Task" [ 965.798853] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.807947] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.808918] env[61006]: DEBUG nova.network.neutron [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 965.896994] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9ac610b-1dd4-4e20-8102-0c8e781b2584 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.853s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.930297] env[61006]: DEBUG nova.network.neutron [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.071224] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337397, 'name': ReconfigVM_Task, 'duration_secs': 0.193863} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.071224] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 966.071224] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d467cd1e-46c3-45e4-8172-485193c49d8a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.095246] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.095895] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-702de8bb-cb66-4b63-bbf8-9fd85d65525e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.119861] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 966.119861] env[61006]: value = "task-1337400" [ 966.119861] env[61006]: _type = "Task" [ 966.119861] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.129801] env[61006]: DEBUG nova.scheduler.client.report [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.133175] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337400, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.213892] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.216185] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.216185] env[61006]: DEBUG nova.network.neutron [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.259022] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337398, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.312929] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337399, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.436752] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Releasing lock "refresh_cache-a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.436752] env[61006]: DEBUG nova.compute.manager [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.436752] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.436752] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356fafa5-f92c-48a0-8b86-a6754872f1f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.464405] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.465127] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aeedb7ca-1e90-4d8a-9c47-2c106891f96d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.475026] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 966.475026] env[61006]: value = "task-1337401" [ 966.475026] env[61006]: _type = "Task" [ 966.475026] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.484747] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.634281] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337400, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.634596] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.635255] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.637961] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.397s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.664331] env[61006]: DEBUG nova.network.neutron [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updated VIF entry in instance network info cache for port 9230a732-200c-4084-8c6c-a5892e9a50ba. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.664331] env[61006]: DEBUG nova.network.neutron [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [{"id": "9230a732-200c-4084-8c6c-a5892e9a50ba", "address": "fa:16:3e:de:76:3b", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9230a732-20", "ovs_interfaceid": "9230a732-200c-4084-8c6c-a5892e9a50ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.760312] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337398, 'name': CreateVM_Task, 'duration_secs': 0.759971} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.760517] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.761345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.761892] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.762408] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.762868] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e053f4ac-de2e-4d52-8677-cd38ac76d159 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.770132] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 966.770132] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5212317a-b67a-1d5b-4d0f-38018a3227e8" [ 966.770132] env[61006]: _type = "Task" [ 966.770132] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.780567] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5212317a-b67a-1d5b-4d0f-38018a3227e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.815182] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337399, 'name': ReconfigVM_Task, 'duration_secs': 0.805251} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.817150] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 0b2b6d52-8340-4f85-9fc0-ab1f2b079031/0b2b6d52-8340-4f85-9fc0-ab1f2b079031.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.818091] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a13f214-a5a2-48d7-813e-ef35d94fe1ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.829016] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 966.829016] env[61006]: value = "task-1337402" [ 966.829016] env[61006]: _type = "Task" [ 966.829016] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.838772] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337402, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.991974] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337401, 'name': PowerOffVM_Task, 'duration_secs': 0.280585} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.992599] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.993176] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.993548] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f21b8822-7d9e-4b42-b8f8-fc77acb2b0a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.024402] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.024837] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.025171] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Deleting the datastore file [datastore1] a1458c3b-f6d4-4cde-ad9a-4e4693ff312c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.025880] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72f8056a-6138-49a5-86ff-22d9910042c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.033593] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for the task: (returnval){ [ 967.033593] env[61006]: value = "task-1337404" [ 967.033593] env[61006]: _type = "Task" [ 967.033593] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.044688] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.117427] env[61006]: DEBUG nova.network.neutron [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.135114] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337400, 'name': ReconfigVM_Task, 'duration_secs': 0.518861} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.135712] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca/6de1f827-1a19-457b-8600-6546593e55ca.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.136268] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 50 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 967.141287] env[61006]: DEBUG nova.compute.utils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.143799] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 967.143799] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 967.147220] env[61006]: INFO nova.compute.claims [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.169034] env[61006]: DEBUG oslo_concurrency.lockutils [req-da9a505d-40ab-4d44-9213-53b99b01c935 req-f1db7e7d-a609-4850-9876-486042f21dd9 service nova] Releasing lock "refresh_cache-66c93148-b91a-4d22-84af-f410c8e10875" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.192932] env[61006]: DEBUG nova.policy [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1932800f85eb4482b1d8774ca048a9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f9a0b972af54498a8f14b9ddbb556d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.280324] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.281279] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Processing image 87693d49-a593-43f4-8a3e-5dc9080f7010 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.284017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.284017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.284017] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.284017] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4aad53b-4663-4d12-b58f-eaac7db8b3a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.292719] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.293114] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.293961] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1f4c0c9-029b-418a-86e1-7b37f80b3da4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.302159] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 967.302159] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5279402f-e174-2668-2abc-fd2153a36cfd" [ 967.302159] env[61006]: _type = "Task" [ 967.302159] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.314671] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5279402f-e174-2668-2abc-fd2153a36cfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.337517] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337402, 'name': Rename_Task, 'duration_secs': 0.350069} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.338160] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.338553] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d36bda01-e785-4fcb-a04a-b6563e4500b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.348266] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 967.348266] env[61006]: value = "task-1337405" [ 967.348266] env[61006]: _type = "Task" [ 967.348266] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.361470] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337405, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.444341] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.444652] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.545517] env[61006]: DEBUG oslo_vmware.api [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Task: {'id': task-1337404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342051} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.546851] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.546851] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.546851] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.546851] env[61006]: INFO nova.compute.manager [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 967.546851] env[61006]: DEBUG oslo.service.loopingcall [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.547268] env[61006]: DEBUG nova.compute.manager [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.547331] env[61006]: DEBUG nova.network.neutron [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 967.587371] env[61006]: DEBUG nova.network.neutron [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.619229] env[61006]: DEBUG oslo_concurrency.lockutils [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.648179] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.652026] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04c2a1b-7477-4796-a6e0-829c39e6b96e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.656565] env[61006]: INFO nova.compute.resource_tracker [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating resource usage from migration 690dbea6-80cc-4e32-9907-818643d640d4 [ 967.679075] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee66db8-f3c1-4a4b-9279-ebb85c7a5849 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.699556] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 67 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 967.740237] env[61006]: DEBUG nova.compute.manager [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.740237] env[61006]: DEBUG nova.compute.manager [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing instance network info cache due to event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.741778] env[61006]: DEBUG oslo_concurrency.lockutils [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.741778] env[61006]: DEBUG oslo_concurrency.lockutils [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.741778] env[61006]: DEBUG nova.network.neutron [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.786320] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Successfully created port: 52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.814165] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 967.814505] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Fetch image to [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3/OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 967.814814] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Downloading stream optimized image 87693d49-a593-43f4-8a3e-5dc9080f7010 to [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3/OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3.vmdk on the data store datastore2 as vApp {{(pid=61006) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 967.814931] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Downloading image file data 87693d49-a593-43f4-8a3e-5dc9080f7010 to the ESX as VM named 'OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3' {{(pid=61006) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 967.881953] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337405, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.919582] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 967.919582] env[61006]: value = "resgroup-9" [ 967.919582] env[61006]: _type = "ResourcePool" [ 967.919582] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 967.919929] env[61006]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-27acbc0a-7909-4812-b2fe-eff8a48496df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.944293] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease: (returnval){ [ 967.944293] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 967.944293] env[61006]: _type = "HttpNfcLease" [ 967.944293] env[61006]: } obtained for vApp import into resource pool (val){ [ 967.944293] env[61006]: value = "resgroup-9" [ 967.944293] env[61006]: _type = "ResourcePool" [ 967.944293] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 967.944783] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the lease: (returnval){ [ 967.944783] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 967.944783] env[61006]: _type = "HttpNfcLease" [ 967.944783] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 967.952371] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.952565] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.952804] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 967.952804] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 967.952804] env[61006]: _type = "HttpNfcLease" [ 967.952804] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 967.957238] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e5f059-cac4-4552-a686-d14bb40536d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.975805] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3158b1be-fab9-405d-b584-69c08e3fe8f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.014834] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfiguring VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 968.018125] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83d78859-c7ab-4075-abd8-68930371b274 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.039322] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 968.039322] env[61006]: value = "task-1337407" [ 968.039322] env[61006]: _type = "Task" [ 968.039322] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.049096] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.066435] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5938b504-3955-4519-9a72-fb914fc133e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.075295] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab3e9e7-eab5-49c5-83b0-d865cc6bd077 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.106796] env[61006]: DEBUG nova.network.neutron [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.111020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62fceb2-c1ee-45bb-9a1d-381741c7a419 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.116864] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8c7e2e-df09-4d21-b935-4f9cd0e68358 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.136947] env[61006]: DEBUG nova.compute.provider_tree [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.156561] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b405ee1d-aa09-43e2-a9c5-36e0fbe227d1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.189630] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbda529a-d122-4ddc-b4cd-5431c14dc5f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.197330] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 83 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 968.273865] env[61006]: DEBUG nova.network.neutron [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Port de9a3378-0e2c-485b-842b-d44c829bee0f binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 968.378842] env[61006]: DEBUG oslo_vmware.api [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337405, 'name': PowerOnVM_Task, 'duration_secs': 0.5634} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.380045] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.380045] env[61006]: INFO nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Took 9.22 seconds to spawn the instance on the hypervisor. [ 968.380045] env[61006]: DEBUG nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.380314] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03c197d-d020-40d3-b2b4-e12bfc1c430c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.453867] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.453867] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 968.453867] env[61006]: _type = "HttpNfcLease" [ 968.453867] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.521518] env[61006]: DEBUG nova.network.neutron [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updated VIF entry in instance network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.521937] env[61006]: DEBUG nova.network.neutron [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.548695] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.612579] env[61006]: INFO nova.compute.manager [-] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Took 1.07 seconds to deallocate network for instance. [ 968.641016] env[61006]: DEBUG nova.scheduler.client.report [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.670688] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.697659] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.697905] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.698078] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.698270] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.698424] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.698573] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.698779] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.698940] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.699128] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.699293] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.699471] env[61006]: DEBUG nova.virt.hardware [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.700340] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992938ba-909a-4668-84f6-b924b8dd6cc1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.704256] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-4c78fb13-f7fd-4240-9ec6-bbad8049dcdd tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance '524f3fd1-1e71-40c0-96c2-0acac5055e01' progress to 100 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 968.713036] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba8de46-8b40-4f14-a19c-483e0ecf16f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.895399] env[61006]: INFO nova.compute.manager [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Took 19.95 seconds to build instance. [ 968.953753] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.953753] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 968.953753] env[61006]: _type = "HttpNfcLease" [ 968.953753] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 968.954058] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 968.954058] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b0bed9-9868-7aef-09ea-a9cc1bea8a9e" [ 968.954058] env[61006]: _type = "HttpNfcLease" [ 968.954058] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 968.954665] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c931ba-63eb-465d-9800-4105847a6e79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.963457] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 968.963637] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 969.024905] env[61006]: DEBUG oslo_concurrency.lockutils [req-33fee583-a0fd-42dc-82b1-c33e531d2fb9 req-c000c0d9-e65b-43b6-b687-9c5e76c2faff service nova] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.031028] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-97c3a3ea-5a8f-4308-a97a-984f8ca57f24 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.049040] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.119454] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.146742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.508s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.146742] env[61006]: INFO nova.compute.manager [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Migrating [ 969.152957] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.664s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.293957] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.295317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.295317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.325161] env[61006]: DEBUG nova.compute.manager [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Received event network-vif-plugged-52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.325390] env[61006]: DEBUG oslo_concurrency.lockutils [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] Acquiring lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.325603] env[61006]: DEBUG oslo_concurrency.lockutils [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.325770] env[61006]: DEBUG oslo_concurrency.lockutils [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.325936] env[61006]: DEBUG nova.compute.manager [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] No waiting events found dispatching network-vif-plugged-52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.327664] env[61006]: WARNING nova.compute.manager [req-10afa7cc-0a9c-4f8c-a583-bd5f618c5383 req-4a7fc271-db90-410b-b6c0-d6929a477ce7 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Received unexpected event network-vif-plugged-52802f58-a9a0-4d6b-a555-cc9e441ea74b for instance with vm_state building and task_state spawning. [ 969.402781] env[61006]: DEBUG oslo_concurrency.lockutils [None req-726004d1-c71d-437c-b11f-453a07109f58 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.466s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.402781] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Successfully updated port: 52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.551564] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.667680] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.667846] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.668025] env[61006]: DEBUG nova.network.neutron [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.768574] env[61006]: DEBUG nova.compute.manager [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.768770] env[61006]: DEBUG nova.compute.manager [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing instance network info cache due to event network-changed-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.768979] env[61006]: DEBUG oslo_concurrency.lockutils [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] Acquiring lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.769137] env[61006]: DEBUG oslo_concurrency.lockutils [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] Acquired lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.769306] env[61006]: DEBUG nova.network.neutron [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Refreshing network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.906591] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.906773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquired lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.906989] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.052576] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.170633] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Applying migration context for instance 524f3fd1-1e71-40c0-96c2-0acac5055e01 as it has an incoming, in-progress migration fa68ecfe-b11e-4599-a1c8-6fbe885ad3d3. Migration status is finished {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 970.170958] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Applying migration context for instance 6de1f827-1a19-457b-8600-6546593e55ca as it has an incoming, in-progress migration 8e633730-59d9-4a02-be82-40f008be263d. Migration status is post-migrating {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 970.171184] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Applying migration context for instance d027fd07-c409-46e5-857f-cdd2c0479f53 as it has an incoming, in-progress migration 690dbea6-80cc-4e32-9907-818643d640d4. Migration status is pre-migrating {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 970.173051] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating resource usage from migration fa68ecfe-b11e-4599-a1c8-6fbe885ad3d3 [ 970.173418] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating resource usage from migration 8e633730-59d9-4a02-be82-40f008be263d [ 970.173739] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating resource usage from migration 690dbea6-80cc-4e32-9907-818643d640d4 [ 970.200321] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0111452e-1b4f-499c-932d-f31364d1a14c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.200491] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance aedaa5d6-e0f2-492c-a14b-3254863e1f06 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.200610] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.200720] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 1a546f17-2fb8-4b99-9001-98cc6fe76837 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.200821] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.200934] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance f048bd9c-048b-4439-bbbf-dfcee7f18d84 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201056] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 02408ca9-c580-444a-9608-a752146ca499 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201198] env[61006]: WARNING nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance a1458c3b-f6d4-4cde-ad9a-4e4693ff312c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 970.201312] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 66c93148-b91a-4d22-84af-f410c8e10875 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201424] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Migration fa68ecfe-b11e-4599-a1c8-6fbe885ad3d3 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 970.201533] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 524f3fd1-1e71-40c0-96c2-0acac5055e01 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201641] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 0b2b6d52-8340-4f85-9fc0-ab1f2b079031 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201747] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Migration 8e633730-59d9-4a02-be82-40f008be263d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 970.201852] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6de1f827-1a19-457b-8600-6546593e55ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.201956] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 3a3aa93a-818e-41d0-896e-72ba97b8b377 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.202120] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Migration 690dbea6-80cc-4e32-9907-818643d640d4 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 970.202212] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d027fd07-c409-46e5-857f-cdd2c0479f53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.380087] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.380287] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.380553] env[61006]: DEBUG nova.network.neutron [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.432477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "095f85df-7265-46a9-bdbb-963292384493" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.432847] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.467106] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.497264] env[61006]: DEBUG nova.network.neutron [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.551779] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.603145] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 970.603558] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 970.604566] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f4c5d6-0cba-4660-bacd-73ad6867a178 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.614540] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 970.614745] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 970.615031] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d8fc68df-b632-4ae8-bffa-21eea3592ffb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.704975] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 095f85df-7265-46a9-bdbb-963292384493 has allocations against this compute host but is not found in the database. [ 970.705236] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 970.705383] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3776MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 970.743281] env[61006]: DEBUG nova.network.neutron [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updated VIF entry in instance network info cache for port 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.743748] env[61006]: DEBUG nova.network.neutron [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [{"id": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "address": "fa:16:3e:1e:54:db", "network": {"id": "3b4f8575-21f5-4bad-8de8-aa438e0f9c99", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1349305243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbbb1cf0f7974ccdb451078aaa448272", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c495ff-c9", "ovs_interfaceid": "43c495ff-c9d9-46fd-93e5-5c92d5f01d1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.745731] env[61006]: DEBUG nova.network.neutron [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Updating instance_info_cache with network_info: [{"id": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "address": "fa:16:3e:b2:04:11", "network": {"id": "d9513017-48d8-4c98-b4d5-28168bdbbd8a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1898333854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f9a0b972af54498a8f14b9ddbb556d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52802f58-a9", "ovs_interfaceid": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.843091] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.843368] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.843591] env[61006]: DEBUG nova.compute.manager [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Going to confirm migration 2 {{(pid=61006) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 970.928470] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1927bde-7455-4834-99a4-7b663a17a0b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.935458] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.938596] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc62420-5d13-4bb0-be4c-852358dca983 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.971060] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d9452f-93f3-4694-bb46-60767df72dad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.980633] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da26c0c-b800-40b1-896e-9ef5cc6cbdff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.996287] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.000576] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.038048] env[61006]: DEBUG oslo_vmware.rw_handles [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52459457-7211-6859-0d0a-af3986ad95aa/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 971.038048] env[61006]: INFO nova.virt.vmwareapi.images [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Downloaded image file data 87693d49-a593-43f4-8a3e-5dc9080f7010 [ 971.038048] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc1e8fc-2b69-46b1-87a8-67f0668d5813 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.061367] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bca6b2f-c21b-4e9f-b1c8-dfc6f4b1c453 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.066567] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.175265] env[61006]: DEBUG nova.network.neutron [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.209784] env[61006]: INFO nova.virt.vmwareapi.images [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] The imported VM was unregistered [ 971.212225] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 971.212528] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.212801] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-573d0391-927b-4a98-a493-1e1ac7c559c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.239351] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010 {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.239619] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3/OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3.vmdk to [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk. {{(pid=61006) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 971.239908] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8050a08d-ebc1-4ca8-8f06-8f45c3394a54 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.246793] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 971.246793] env[61006]: value = "task-1337409" [ 971.246793] env[61006]: _type = "Task" [ 971.246793] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.247294] env[61006]: DEBUG oslo_concurrency.lockutils [req-d091882b-161c-4455-a5e8-10d54026dcab req-e6871229-0e7f-49e2-a3c1-fac2b7bb4dbd service nova] Releasing lock "refresh_cache-1a546f17-2fb8-4b99-9001-98cc6fe76837" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.250460] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Releasing lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.250734] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Instance network_info: |[{"id": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "address": "fa:16:3e:b2:04:11", "network": {"id": "d9513017-48d8-4c98-b4d5-28168bdbbd8a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1898333854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f9a0b972af54498a8f14b9ddbb556d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52802f58-a9", "ovs_interfaceid": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 971.251395] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:04:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52802f58-a9a0-4d6b-a555-cc9e441ea74b', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.258706] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Creating folder: Project (1f9a0b972af54498a8f14b9ddbb556d7). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.259288] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8c1ff2b-670b-43cc-bd89-9496a59ca1c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.263774] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.271795] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Created folder: Project (1f9a0b972af54498a8f14b9ddbb556d7) in parent group-v285275. [ 971.271979] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Creating folder: Instances. Parent ref: group-v285411. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.272238] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0cac376-b6ce-4013-8528-6f74fd8c7cfb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.281039] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Created folder: Instances in parent group-v285411. [ 971.281039] env[61006]: DEBUG oslo.service.loopingcall [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.281039] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.281039] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ab05b00-ec0a-43cd-8ac5-6b34a05583d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.301958] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.301958] env[61006]: value = "task-1337412" [ 971.301958] env[61006]: _type = "Task" [ 971.301958] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.309131] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337412, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.355135] env[61006]: DEBUG nova.compute.manager [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Received event network-changed-52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.355346] env[61006]: DEBUG nova.compute.manager [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Refreshing instance network info cache due to event network-changed-52802f58-a9a0-4d6b-a555-cc9e441ea74b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 971.355561] env[61006]: DEBUG oslo_concurrency.lockutils [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] Acquiring lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.355701] env[61006]: DEBUG oslo_concurrency.lockutils [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] Acquired lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.355863] env[61006]: DEBUG nova.network.neutron [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Refreshing network info cache for port 52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.384344] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.384557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.384737] env[61006]: DEBUG nova.network.neutron [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.384920] env[61006]: DEBUG nova.objects.instance [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'info_cache' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.460557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.497663] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.553991] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.678612] env[61006]: DEBUG oslo_concurrency.lockutils [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.756921] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.811616] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337412, 'name': CreateVM_Task, 'duration_secs': 0.50649} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.811838] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.812568] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.812779] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.813083] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.813364] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27790fc2-a774-4848-a5b2-1773c3636b08 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.818081] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 971.818081] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ba53b8-b51b-6df0-4db1-eaab7105a3ba" [ 971.818081] env[61006]: _type = "Task" [ 971.818081] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.825845] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ba53b8-b51b-6df0-4db1-eaab7105a3ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.003407] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 972.003639] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.851s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.003946] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.210s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.004152] env[61006]: DEBUG nova.objects.instance [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61006) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 972.056630] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.085270] env[61006]: DEBUG nova.network.neutron [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Updated VIF entry in instance network info cache for port 52802f58-a9a0-4d6b-a555-cc9e441ea74b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.085645] env[61006]: DEBUG nova.network.neutron [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Updating instance_info_cache with network_info: [{"id": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "address": "fa:16:3e:b2:04:11", "network": {"id": "d9513017-48d8-4c98-b4d5-28168bdbbd8a", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1898333854-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f9a0b972af54498a8f14b9ddbb556d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52802f58-a9", "ovs_interfaceid": "52802f58-a9a0-4d6b-a555-cc9e441ea74b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.201860] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498f3996-6bc8-40ef-aca5-1030dd6be456 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.223020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7561bca0-38c1-4028-bbb3-2186eaf751c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.230667] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 83 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.261081] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.328295] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ba53b8-b51b-6df0-4db1-eaab7105a3ba, 'name': SearchDatastore_Task, 'duration_secs': 0.017543} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.328646] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.328891] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.329162] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.329317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.329502] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.329782] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0d3b3e9-77c3-4459-abd7-26966d0fe97d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.339346] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.339524] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.340310] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-610ef52b-e6b7-40ac-b6b5-0e4e60f7c15c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.346758] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 972.346758] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e9a95d-ce67-0e11-2f12-460dfbe5aeee" [ 972.346758] env[61006]: _type = "Task" [ 972.346758] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.355188] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e9a95d-ce67-0e11-2f12-460dfbe5aeee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.518740] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e76b9c4-a310-46d6-92ad-2149768f514e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.540398] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 0 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.553843] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.589539] env[61006]: DEBUG oslo_concurrency.lockutils [req-52b82fdc-afb4-4630-b371-7bbf08c4a82f req-cd152207-3645-455b-848e-8353627d3bc5 service nova] Releasing lock "refresh_cache-3a3aa93a-818e-41d0-896e-72ba97b8b377" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.646909] env[61006]: DEBUG nova.network.neutron [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.736817] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.737214] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8278835c-1a8d-426a-8527-0a55e61f866a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.745668] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 972.745668] env[61006]: value = "task-1337413" [ 972.745668] env[61006]: _type = "Task" [ 972.745668] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.758632] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.762489] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.858877] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e9a95d-ce67-0e11-2f12-460dfbe5aeee, 'name': SearchDatastore_Task, 'duration_secs': 0.015985} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.859716] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b22c40ab-008f-497c-b51e-9d8f28b9b4c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.865770] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 972.865770] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52369410-d61c-24ab-8d5e-77d74546a31b" [ 972.865770] env[61006]: _type = "Task" [ 972.865770] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.875532] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52369410-d61c-24ab-8d5e-77d74546a31b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.018352] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a1140337-97f9-4031-9ec2-ed104070c90a tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.019737] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.900s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.019979] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.022353] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.562s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.024033] env[61006]: INFO nova.compute.claims [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 973.046577] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.048073] env[61006]: INFO nova.scheduler.client.report [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Deleted allocations for instance a1458c3b-f6d4-4cde-ad9a-4e4693ff312c [ 973.050973] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9765fce8-73e9-481e-8e98-a33e0e32994c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.066718] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.069322] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 973.069322] env[61006]: value = "task-1337414" [ 973.069322] env[61006]: _type = "Task" [ 973.069322] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.079803] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.149287] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.149678] env[61006]: DEBUG nova.objects.instance [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'migration_context' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.259533] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337413, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.263328] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.377270] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52369410-d61c-24ab-8d5e-77d74546a31b, 'name': SearchDatastore_Task, 'duration_secs': 0.02249} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.377714] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.377959] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 3a3aa93a-818e-41d0-896e-72ba97b8b377/3a3aa93a-818e-41d0-896e-72ba97b8b377.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.378264] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63cd3213-a79f-4cc9-b26b-ed779fa8ca98 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.385856] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 973.385856] env[61006]: value = "task-1337415" [ 973.385856] env[61006]: _type = "Task" [ 973.385856] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.394696] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337415, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.558542] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.562538] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bf113539-7d27-4823-8375-81d6ed47d6b4 tempest-ServerShowV254Test-1339604622 tempest-ServerShowV254Test-1339604622-project-member] Lock "a1458c3b-f6d4-4cde-ad9a-4e4693ff312c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.316s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.580451] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337414, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.653608] env[61006]: DEBUG nova.objects.base [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Object Instance<524f3fd1-1e71-40c0-96c2-0acac5055e01> lazy-loaded attributes: info_cache,migration_context {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 973.655047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cdc593-d605-48e8-b9d3-ef80fcfe870d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.675838] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0f1d51b-2502-4cb7-a877-d1cefb9559a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.683272] env[61006]: DEBUG oslo_vmware.api [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 973.683272] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ab67be-cacd-6485-d401-7cf31b100cd3" [ 973.683272] env[61006]: _type = "Task" [ 973.683272] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.694295] env[61006]: DEBUG oslo_vmware.api [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ab67be-cacd-6485-d401-7cf31b100cd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.760245] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337413, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.763920] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.895860] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337415, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.063940] env[61006]: DEBUG oslo_vmware.api [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337407, 'name': ReconfigVM_Task, 'duration_secs': 5.841469} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.064643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.064937] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Reconfigured VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 974.080136] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337414, 'name': PowerOffVM_Task, 'duration_secs': 0.538518} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.083395] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.083738] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 17 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.197649] env[61006]: DEBUG oslo_vmware.api [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ab67be-cacd-6485-d401-7cf31b100cd3, 'name': SearchDatastore_Task, 'duration_secs': 0.048691} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.198037] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.264676] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337413, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.270876] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337409, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.776325} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.271508] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3/OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3.vmdk to [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk. [ 974.271719] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Cleaning up location [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 974.271952] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_236f6680-b437-478d-b5f4-ec098c2af1b3 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.272183] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44cd2001-90fc-469c-9ff0-920535631330 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.279930] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 974.279930] env[61006]: value = "task-1337416" [ 974.279930] env[61006]: _type = "Task" [ 974.279930] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.291558] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.308806] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93124027-0f1b-49c9-a963-cbbc04a2d7ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.316096] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c282c78-a723-4cbd-a972-e0ddad1898e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.348397] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb5cb3d-b8ee-42d9-a160-dada5d8bf990 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.356833] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a762af-2b9e-4d65-8dc7-c5b2dd1692ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.370873] env[61006]: DEBUG nova.compute.provider_tree [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.396680] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337415, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.591530] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.592116] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.592116] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.592116] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.592492] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.592492] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.592564] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.592959] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.592959] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.593131] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.593373] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.599063] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da610c03-cb6f-4133-b116-3a324bd693ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.616398] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 974.616398] env[61006]: value = "task-1337417" [ 974.616398] env[61006]: _type = "Task" [ 974.616398] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.624574] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337417, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.761030] env[61006]: DEBUG oslo_vmware.api [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337413, 'name': PowerOnVM_Task, 'duration_secs': 1.542296} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.761030] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.761466] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-82a6502d-3a06-48d9-a9f5-2f62f8d60974 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance '6de1f827-1a19-457b-8600-6546593e55ca' progress to 100 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.789770] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196067} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.790108] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.790279] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.790564] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk to [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.790845] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63a9858d-ae78-4147-b06c-681eff835710 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.797222] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 974.797222] env[61006]: value = "task-1337418" [ 974.797222] env[61006]: _type = "Task" [ 974.797222] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.805182] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.874758] env[61006]: DEBUG nova.scheduler.client.report [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.899021] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337415, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.016297} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.899298] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] 3a3aa93a-818e-41d0-896e-72ba97b8b377/3a3aa93a-818e-41d0-896e-72ba97b8b377.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.899520] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.900038] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a07186aa-a8c7-4285-a341-552d647f00ac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.907157] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 974.907157] env[61006]: value = "task-1337419" [ 974.907157] env[61006]: _type = "Task" [ 974.907157] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.915502] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.128072] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337417, 'name': ReconfigVM_Task, 'duration_secs': 0.361019} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.128335] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 33 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.306667] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.380602] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.381110] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 975.383941] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.186s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.417752] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.418019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.418898] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb368ff7-1aa2-4b8c-99dc-89514169fa6e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.435129] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.435320] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.435501] env[61006]: DEBUG nova.network.neutron [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.445173] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 3a3aa93a-818e-41d0-896e-72ba97b8b377/3a3aa93a-818e-41d0-896e-72ba97b8b377.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.447960] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d218df6e-16f6-44ae-a182-a8e3965d6aef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.468897] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 975.468897] env[61006]: value = "task-1337420" [ 975.468897] env[61006]: _type = "Task" [ 975.468897] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.477173] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.635169] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.635608] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.635608] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.635608] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.635719] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.635892] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.636071] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.636241] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.636412] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.636580] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.636752] env[61006]: DEBUG nova.virt.hardware [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.642134] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.642490] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a0e0b0f-0e53-4ceb-9c95-a8f92291b764 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.663297] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 975.663297] env[61006]: value = "task-1337421" [ 975.663297] env[61006]: _type = "Task" [ 975.663297] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.671455] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337421, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.735817] env[61006]: INFO nova.network.neutron [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Port f07eb7f2-10e5-4957-a5c8-3792f6d15b26 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 975.736251] env[61006]: DEBUG nova.network.neutron [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.808319] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.888038] env[61006]: DEBUG nova.compute.utils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.893267] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.893267] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.968478] env[61006]: DEBUG nova.policy [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 976.011778] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.175576] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337421, 'name': ReconfigVM_Task, 'duration_secs': 0.394042} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.175905] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.179445] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3c1b8b-6ad8-4538-83af-3cd8cc06f9b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.204373] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.207163] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1812b13a-bbb8-4db9-b63e-298ff2525a56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.230165] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 976.230165] env[61006]: value = "task-1337422" [ 976.230165] env[61006]: _type = "Task" [ 976.230165] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.236172] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a57196-b4b9-482a-836e-381eae3addec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.243947] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.246297] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337422, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.251383] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22517537-841c-4372-a0f0-97ec3dfc5eb1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.294642] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db64023c-fd81-463c-bab3-927ea24cd27e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.311423] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b355f6e-af33-4a9a-b4b4-9b3127222ec3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.320140] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.337434] env[61006]: DEBUG nova.compute.provider_tree [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.385876] env[61006]: DEBUG nova.compute.manager [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.386113] env[61006]: DEBUG nova.compute.manager [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing instance network info cache due to event network-changed-58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.386369] env[61006]: DEBUG oslo_concurrency.lockutils [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] Acquiring lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.386518] env[61006]: DEBUG oslo_concurrency.lockutils [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] Acquired lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.386678] env[61006]: DEBUG nova.network.neutron [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Refreshing network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.394181] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 976.448853] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Successfully created port: 45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.481365] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337420, 'name': ReconfigVM_Task, 'duration_secs': 0.571212} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.481671] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 3a3aa93a-818e-41d0-896e-72ba97b8b377/3a3aa93a-818e-41d0-896e-72ba97b8b377.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.482447] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8613b2b-03ac-4acd-9c7e-409785fa5989 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.490341] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 976.490341] env[61006]: value = "task-1337423" [ 976.490341] env[61006]: _type = "Task" [ 976.490341] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.501339] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337423, 'name': Rename_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.740662] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.748872] env[61006]: DEBUG oslo_concurrency.lockutils [None req-04b1b08c-8940-4521-9538-8d71414d6564 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.304s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.810970] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.841347] env[61006]: DEBUG nova.scheduler.client.report [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.859764] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.860054] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.860264] env[61006]: DEBUG nova.compute.manager [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Going to confirm migration 3 {{(pid=61006) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 977.003086] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337423, 'name': Rename_Task, 'duration_secs': 0.202738} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.005729] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.006023] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75fc1a93-6b04-4aa8-9be4-9bb6d900f217 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.013062] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 977.013062] env[61006]: value = "task-1337424" [ 977.013062] env[61006]: _type = "Task" [ 977.013062] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.020745] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.074652] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.074950] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.075354] env[61006]: DEBUG nova.objects.instance [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'flavor' on Instance uuid f048bd9c-048b-4439-bbbf-dfcee7f18d84 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.165206] env[61006]: DEBUG nova.network.neutron [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updated VIF entry in instance network info cache for port 58bb0db5-4669-4185-8d20-b5a77724df9b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 977.165615] env[61006]: DEBUG nova.network.neutron [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [{"id": "58bb0db5-4669-4185-8d20-b5a77724df9b", "address": "fa:16:3e:aa:9e:5d", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58bb0db5-46", "ovs_interfaceid": "58bb0db5-4669-4185-8d20-b5a77724df9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.242835] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.311699] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.406182] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 977.434333] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.434333] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.434333] env[61006]: DEBUG nova.network.neutron [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.434333] env[61006]: DEBUG nova.objects.instance [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lazy-loading 'info_cache' on Instance uuid 6de1f827-1a19-457b-8600-6546593e55ca {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.439025] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.439282] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.439442] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.439629] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.439778] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.439916] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.440380] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.440576] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.440764] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.441017] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.441289] env[61006]: DEBUG nova.virt.hardware [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.442378] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83d43ea-23c0-4ac3-9d71-5c393ca6c1bc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.451998] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450a810a-d34e-4df8-8f73-e6ddb881e149 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.523762] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337424, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.668460] env[61006]: DEBUG oslo_concurrency.lockutils [req-4d78c049-7a8d-47ad-91f7-3d614ae63d27 req-213de77b-155d-4743-b907-d60fa8810f84 service nova] Releasing lock "refresh_cache-bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.741722] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.813296] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.855025] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.471s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.855536] env[61006]: DEBUG nova.compute.manager [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61006) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 978.014913] env[61006]: DEBUG nova.objects.instance [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'pci_requests' on Instance uuid f048bd9c-048b-4439-bbbf-dfcee7f18d84 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.025343] env[61006]: DEBUG oslo_vmware.api [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337424, 'name': PowerOnVM_Task, 'duration_secs': 0.759424} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.025654] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.026051] env[61006]: INFO nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Took 9.35 seconds to spawn the instance on the hypervisor. [ 978.026051] env[61006]: DEBUG nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.026785] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adcb250-e7be-4f9b-b7f5-c040afc2016f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.240734] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337422, 'name': ReconfigVM_Task, 'duration_secs': 1.570492} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.241009] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Reconfigured VM instance instance-00000060 to attach disk [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53/d027fd07-c409-46e5-857f-cdd2c0479f53.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.241297] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 50 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.308647] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Successfully updated port: 45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 978.315973] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337418, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.096961} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.316245] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/87693d49-a593-43f4-8a3e-5dc9080f7010/87693d49-a593-43f4-8a3e-5dc9080f7010.vmdk to [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.317085] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15c4ee0-0949-4e2c-9836-580d5e5efd50 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.339319] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.340589] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee6cda0d-20ca-4fb1-b7e7-cebea9022768 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.355480] env[61006]: DEBUG nova.compute.manager [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Received event network-vif-plugged-45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.355685] env[61006]: DEBUG oslo_concurrency.lockutils [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] Acquiring lock "095f85df-7265-46a9-bdbb-963292384493-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.355890] env[61006]: DEBUG oslo_concurrency.lockutils [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] Lock "095f85df-7265-46a9-bdbb-963292384493-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.356078] env[61006]: DEBUG oslo_concurrency.lockutils [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] Lock "095f85df-7265-46a9-bdbb-963292384493-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.356611] env[61006]: DEBUG nova.compute.manager [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] No waiting events found dispatching network-vif-plugged-45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.356611] env[61006]: WARNING nova.compute.manager [req-1faffa4c-1564-4538-91f4-f4af36d7c1bf req-43dafdac-96c9-4ebe-9634-cdd8f9ee9c50 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Received unexpected event network-vif-plugged-45c0cda2-36f4-4b9a-9e88-07b136107358 for instance with vm_state building and task_state spawning. [ 978.363417] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 978.363417] env[61006]: value = "task-1337425" [ 978.363417] env[61006]: _type = "Task" [ 978.363417] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.374229] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337425, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.414954] env[61006]: DEBUG nova.compute.manager [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.415536] env[61006]: DEBUG nova.compute.manager [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing instance network info cache due to event network-changed-5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.415862] env[61006]: DEBUG oslo_concurrency.lockutils [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.416096] env[61006]: DEBUG oslo_concurrency.lockutils [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.416332] env[61006]: DEBUG nova.network.neutron [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.425157] env[61006]: INFO nova.scheduler.client.report [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted allocation for migration fa68ecfe-b11e-4599-a1c8-6fbe885ad3d3 [ 978.520450] env[61006]: DEBUG nova.objects.base [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 978.520728] env[61006]: DEBUG nova.network.neutron [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.547470] env[61006]: INFO nova.compute.manager [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Took 17.53 seconds to build instance. [ 978.669696] env[61006]: DEBUG nova.policy [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bcbdd566bbe04595a475cf805d5a4b2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '85b03e26e0034e30b74761724d0a39e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.748286] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291664e6-578a-466b-920e-0aa9568a4e1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.773556] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8f4702-ee79-4640-a82b-6bb4992b7f51 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.791736] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 67 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.811316] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.811316] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.811495] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.839783] env[61006]: DEBUG nova.network.neutron [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [{"id": "de9a3378-0e2c-485b-842b-d44c829bee0f", "address": "fa:16:3e:2f:e8:f8", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde9a3378-0e", "ovs_interfaceid": "de9a3378-0e2c-485b-842b-d44c829bee0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.875323] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337425, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.934227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cb4047c7-8354-4be1-aafd-eeb3d5a908eb tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.091s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.049702] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c952549-7a97-4e46-a114-a9e3b5caff8c tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.044s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.292520] env[61006]: DEBUG nova.network.neutron [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updated VIF entry in instance network info cache for port 5d945d09-25a9-4087-96d5-f0be90bcddff. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.292978] env[61006]: DEBUG nova.network.neutron [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.340829] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-6de1f827-1a19-457b-8600-6546593e55ca" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.341164] env[61006]: DEBUG nova.objects.instance [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lazy-loading 'migration_context' on Instance uuid 6de1f827-1a19-457b-8600-6546593e55ca {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.348809] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 979.363123] env[61006]: DEBUG nova.network.neutron [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Port 1a7b1ed1-7ded-4bd0-880f-96439cf27513 binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 979.375865] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337425, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.487035] env[61006]: DEBUG nova.network.neutron [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Updating instance_info_cache with network_info: [{"id": "45c0cda2-36f4-4b9a-9e88-07b136107358", "address": "fa:16:3e:fb:00:2b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c0cda2-36", "ovs_interfaceid": "45c0cda2-36f4-4b9a-9e88-07b136107358", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.752684] env[61006]: DEBUG nova.objects.instance [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.796230] env[61006]: DEBUG oslo_concurrency.lockutils [req-4ce10dd0-bd5e-4c9a-aa07-7148097eed9c req-5c1fdafb-db90-4c85-b65b-96d0e9df8fd2 service nova] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.844238] env[61006]: DEBUG nova.objects.base [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Object Instance<6de1f827-1a19-457b-8600-6546593e55ca> lazy-loaded attributes: info_cache,migration_context {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 979.845219] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1878a0-fda3-4aec-a88c-a07877fd95ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.869891] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c3e602a-cdfc-4705-8ac7-c19fc5f55603 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.880024] env[61006]: DEBUG oslo_vmware.api [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 979.880024] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527685ec-8de0-b719-b0ca-6cc8180ef536" [ 979.880024] env[61006]: _type = "Task" [ 979.880024] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.882903] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337425, 'name': ReconfigVM_Task, 'duration_secs': 1.201994} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.885867] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875/66c93148-b91a-4d22-84af-f410c8e10875.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.887463] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'boot_index': 0, 'encrypted': False, 'size': 0, 'disk_bus': None, 'encryption_options': None, 'device_type': 'disk', 'encryption_format': None, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'image_id': '2fb15142-70f7-4f09-b4ca-b3b9a23e35e0'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285400', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'name': 'volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '66c93148-b91a-4d22-84af-f410c8e10875', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'serial': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a'}, 'guest_format': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'attachment_id': 'ccb702b0-5686-4e95-aade-f3524cf84c49', 'device_type': None, 'delete_on_termination': False, 'volume_type': None}], 'swap': None} {{(pid=61006) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 979.887751] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 979.888020] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285400', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'name': 'volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '66c93148-b91a-4d22-84af-f410c8e10875', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'serial': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 979.889318] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b50b8c-90a2-4dd1-9d2c-a35911de31dd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.896897] env[61006]: DEBUG oslo_vmware.api [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]527685ec-8de0-b719-b0ca-6cc8180ef536, 'name': SearchDatastore_Task, 'duration_secs': 0.008614} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.907903] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.908236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.910723] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf518526-4476-4742-a9c8-b41552ca92f8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.934991] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a/volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.936379] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-282d602b-27f9-436f-80dd-e7e5b2030e85 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.955203] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 979.955203] env[61006]: value = "task-1337426" [ 979.955203] env[61006]: _type = "Task" [ 979.955203] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.963359] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.988728] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.989684] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Instance network_info: |[{"id": "45c0cda2-36f4-4b9a-9e88-07b136107358", "address": "fa:16:3e:fb:00:2b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c0cda2-36", "ovs_interfaceid": "45c0cda2-36f4-4b9a-9e88-07b136107358", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.990192] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:00:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45c0cda2-36f4-4b9a-9e88-07b136107358', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.998570] env[61006]: DEBUG oslo.service.loopingcall [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.998805] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.999084] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e06183e6-65d0-4ffa-bc73-0d60b4826765 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.019910] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.019910] env[61006]: value = "task-1337427" [ 980.019910] env[61006]: _type = "Task" [ 980.019910] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.029974] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337427, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.257881] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.258119] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.258303] env[61006]: DEBUG nova.network.neutron [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.258495] env[61006]: DEBUG nova.objects.instance [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'info_cache' on Instance uuid 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.364403] env[61006]: DEBUG nova.compute.manager [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Received event network-changed-45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.364653] env[61006]: DEBUG nova.compute.manager [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Refreshing instance network info cache due to event network-changed-45c0cda2-36f4-4b9a-9e88-07b136107358. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.364907] env[61006]: DEBUG oslo_concurrency.lockutils [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] Acquiring lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.365090] env[61006]: DEBUG oslo_concurrency.lockutils [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] Acquired lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.365257] env[61006]: DEBUG nova.network.neutron [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Refreshing network info cache for port 45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 980.386386] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.386626] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.386802] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.468730] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337426, 'name': ReconfigVM_Task, 'duration_secs': 0.334899} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.469011] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a/volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.481349] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef9f6204-0b1b-4805-a3fa-4a0c63e6a7b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.498424] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 980.498424] env[61006]: value = "task-1337428" [ 980.498424] env[61006]: _type = "Task" [ 980.498424] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.506763] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.531951] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337427, 'name': CreateVM_Task, 'duration_secs': 0.321139} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.532147] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 980.536250] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.536459] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.536798] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.537603] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a62c92d3-c063-4b44-ac62-b1a4f3f14152 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.542315] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 980.542315] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]523dd1f9-c620-7dd0-9547-cc53bd0f9fb3" [ 980.542315] env[61006]: _type = "Task" [ 980.542315] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.551755] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]523dd1f9-c620-7dd0-9547-cc53bd0f9fb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.576302] env[61006]: DEBUG nova.network.neutron [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Successfully updated port: f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.615849] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.616192] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.616580] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.616875] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.617144] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.619406] env[61006]: INFO nova.compute.manager [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Terminating instance [ 980.621530] env[61006]: DEBUG nova.compute.manager [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.621743] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.622715] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e05afd-e2eb-4fe6-a0f7-688f441e4276 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.636029] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.636229] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca2f225c-1ffb-47b5-a701-42de5ecb8de9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.644123] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 980.644123] env[61006]: value = "task-1337429" [ 980.644123] env[61006]: _type = "Task" [ 980.644123] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.653894] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.699133] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07299e91-63b1-47a6-9b4e-cd994145dd49 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.707689] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd9e4b8-fe84-4cdb-a1b5-f81a45f87fe1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.740453] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cf2fcf-29c0-4e41-a4e6-c2e71b488bb1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.748479] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a825ef81-1aca-4401-b71e-277911d1bd2b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.763453] env[61006]: DEBUG nova.objects.base [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Object Instance<524f3fd1-1e71-40c0-96c2-0acac5055e01> lazy-loaded attributes: flavor,info_cache {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 980.764982] env[61006]: DEBUG nova.compute.provider_tree [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.011228] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337428, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.051785] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]523dd1f9-c620-7dd0-9547-cc53bd0f9fb3, 'name': SearchDatastore_Task, 'duration_secs': 0.01128} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.052098] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.052366] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.052608] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.052763] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.053098] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.053234] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-498954bd-180a-4e40-a78f-09c2c8b7dc8d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.070355] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.070612] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 981.071362] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6206e7e-a1f1-4c99-8226-0688209df4ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.075951] env[61006]: DEBUG nova.network.neutron [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Updated VIF entry in instance network info cache for port 45c0cda2-36f4-4b9a-9e88-07b136107358. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 981.076375] env[61006]: DEBUG nova.network.neutron [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Updating instance_info_cache with network_info: [{"id": "45c0cda2-36f4-4b9a-9e88-07b136107358", "address": "fa:16:3e:fb:00:2b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45c0cda2-36", "ovs_interfaceid": "45c0cda2-36f4-4b9a-9e88-07b136107358", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.078796] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 981.078796] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529d28fb-0259-a370-ec18-783563f962c8" [ 981.078796] env[61006]: _type = "Task" [ 981.078796] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.079560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.079727] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.079898] env[61006]: DEBUG nova.network.neutron [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.089972] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]529d28fb-0259-a370-ec18-783563f962c8, 'name': SearchDatastore_Task, 'duration_secs': 0.010889} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.091602] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cdd0b46-a61e-4cb1-9afe-b8c7ac7903bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.097277] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 981.097277] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3a19a-d011-629c-bb6c-8f25e7b98ff0" [ 981.097277] env[61006]: _type = "Task" [ 981.097277] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.104915] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3a19a-d011-629c-bb6c-8f25e7b98ff0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.153541] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337429, 'name': PowerOffVM_Task, 'duration_secs': 0.508232} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.153811] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.153986] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.154259] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2adf55f-f5ec-4515-abac-16e42c6619df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.236017] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.236274] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.236454] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Deleting the datastore file [datastore1] 3a3aa93a-818e-41d0-896e-72ba97b8b377 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.236718] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a3b0644-555e-4069-b428-8ca32aeac875 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.243013] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for the task: (returnval){ [ 981.243013] env[61006]: value = "task-1337431" [ 981.243013] env[61006]: _type = "Task" [ 981.243013] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.250667] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.271618] env[61006]: DEBUG nova.scheduler.client.report [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.426219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.426457] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.426641] env[61006]: DEBUG nova.network.neutron [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.511919] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337428, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.580438] env[61006]: DEBUG oslo_concurrency.lockutils [req-e2ec2c88-ff63-4d57-aac8-5fa54116f92b req-888b69b4-adaa-425c-acd7-f35b26771530 service nova] Releasing lock "refresh_cache-095f85df-7265-46a9-bdbb-963292384493" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.597769] env[61006]: DEBUG nova.network.neutron [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [{"id": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "address": "fa:16:3e:80:9c:f9", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d9ba3b1-a2", "ovs_interfaceid": "3d9ba3b1-a27e-4002-bd16-cfd881fca452", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.610064] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3a19a-d011-629c-bb6c-8f25e7b98ff0, 'name': SearchDatastore_Task, 'duration_secs': 0.016097} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.610907] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.611181] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 095f85df-7265-46a9-bdbb-963292384493/095f85df-7265-46a9-bdbb-963292384493.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.611442] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b3d1c6f-97eb-4dd2-8760-db68adae75c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.619083] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 981.619083] env[61006]: value = "task-1337432" [ 981.619083] env[61006]: _type = "Task" [ 981.619083] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.619853] env[61006]: WARNING nova.network.neutron [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] 98dcba29-9f32-40eb-b526-46e091c2a330 already exists in list: networks containing: ['98dcba29-9f32-40eb-b526-46e091c2a330']. ignoring it [ 981.629101] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.752788] env[61006]: DEBUG oslo_vmware.api [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Task: {'id': task-1337431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176362} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.752981] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.753182] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.753427] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.753634] env[61006]: INFO nova.compute.manager [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Took 1.13 seconds to destroy the instance on the hypervisor. [ 981.753868] env[61006]: DEBUG oslo.service.loopingcall [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.754070] env[61006]: DEBUG nova.compute.manager [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.754165] env[61006]: DEBUG nova.network.neutron [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.908173] env[61006]: DEBUG nova.network.neutron [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "address": "fa:16:3e:0a:c8:a3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07eb7f2-10", "ovs_interfaceid": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.011727] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337428, 'name': ReconfigVM_Task, 'duration_secs': 1.181106} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.012014] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285400', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'name': 'volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '66c93148-b91a-4d22-84af-f410c8e10875', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'serial': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 982.013086] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc893632-dd0e-44c7-a795-3980c69d84bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.022096] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 982.022096] env[61006]: value = "task-1337433" [ 982.022096] env[61006]: _type = "Task" [ 982.022096] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.031387] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337433, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.103207] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-524f3fd1-1e71-40c0-96c2-0acac5055e01" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.137581] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508221} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.139945] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 095f85df-7265-46a9-bdbb-963292384493/095f85df-7265-46a9-bdbb-963292384493.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.140190] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.140692] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8f0e6b6-b184-4d59-98f1-0f3dcfefc5af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.147597] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 982.147597] env[61006]: value = "task-1337434" [ 982.147597] env[61006]: _type = "Task" [ 982.147597] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.156703] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.180838] env[61006]: DEBUG nova.network.neutron [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.283071] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.374s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.396471] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.396471] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.396596] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.396749] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.396971] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] No waiting events found dispatching network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.397187] env[61006]: WARNING nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received unexpected event network-vif-plugged-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 for instance with vm_state active and task_state None. [ 982.397379] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-changed-f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.397559] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing instance network info cache due to event network-changed-f07eb7f2-10e5-4957-a5c8-3792f6d15b26. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.397749] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.410856] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.411567] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.411739] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.412038] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.412245] env[61006]: DEBUG nova.network.neutron [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Refreshing network info cache for port f07eb7f2-10e5-4957-a5c8-3792f6d15b26 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.414956] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cf326a-7bf2-4bf8-a199-daf199c62d30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.434121] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.434369] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.434528] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.434707] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.434853] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.435017] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.435256] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.435420] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.435589] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.435753] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.435923] env[61006]: DEBUG nova.virt.hardware [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.442481] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfiguring VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 982.443244] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13a9f553-09ab-49c5-91a4-758900d246b6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.460755] env[61006]: DEBUG oslo_vmware.api [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 982.460755] env[61006]: value = "task-1337435" [ 982.460755] env[61006]: _type = "Task" [ 982.460755] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.470729] env[61006]: DEBUG oslo_vmware.api [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337435, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.529928] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337433, 'name': Rename_Task, 'duration_secs': 0.216029} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.530207] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.530470] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a766378-b85b-42a7-ac2b-87a623aae044 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.536050] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 982.536050] env[61006]: value = "task-1337436" [ 982.536050] env[61006]: _type = "Task" [ 982.536050] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.544101] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.613578] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.613755] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7bc88eb1-7712-4b1a-abf3-517f0f0701ae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.623512] env[61006]: DEBUG oslo_vmware.api [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 982.623512] env[61006]: value = "task-1337437" [ 982.623512] env[61006]: _type = "Task" [ 982.623512] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.631715] env[61006]: DEBUG nova.network.neutron [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.632959] env[61006]: DEBUG oslo_vmware.api [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337437, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.658060] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066343} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.658353] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.659148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b02c039-6b31-4910-ac7a-38321e758bb3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.681410] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 095f85df-7265-46a9-bdbb-963292384493/095f85df-7265-46a9-bdbb-963292384493.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.682065] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1bdb11d-dd27-41fb-bb4f-fb4b926b524e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.696722] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.705967] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 982.705967] env[61006]: value = "task-1337438" [ 982.705967] env[61006]: _type = "Task" [ 982.705967] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.714794] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.848106] env[61006]: INFO nova.scheduler.client.report [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocation for migration 8e633730-59d9-4a02-be82-40f008be263d [ 982.971810] env[61006]: DEBUG oslo_vmware.api [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.047151] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337436, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.136465] env[61006]: INFO nova.compute.manager [-] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Took 1.38 seconds to deallocate network for instance. [ 983.136814] env[61006]: DEBUG oslo_vmware.api [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337437, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.198767] env[61006]: DEBUG nova.network.neutron [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updated VIF entry in instance network info cache for port f07eb7f2-10e5-4957-a5c8-3792f6d15b26. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.198767] env[61006]: DEBUG nova.network.neutron [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "address": "fa:16:3e:0a:c8:a3", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07eb7f2-10", "ovs_interfaceid": "f07eb7f2-10e5-4957-a5c8-3792f6d15b26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.217521] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337438, 'name': ReconfigVM_Task, 'duration_secs': 0.424833} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.217967] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 095f85df-7265-46a9-bdbb-963292384493/095f85df-7265-46a9-bdbb-963292384493.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.219435] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfc84b18-3645-441c-97d8-3c5c93b9df54 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.221949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc316cb-1092-4326-8e96-5fd6d9ee58ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.242661] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bccc1be-b1d4-4393-8c84-8dbc45a17646 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.245528] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 983.245528] env[61006]: value = "task-1337439" [ 983.245528] env[61006]: _type = "Task" [ 983.245528] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.251979] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 83 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 983.261090] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337439, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.353731] env[61006]: DEBUG oslo_concurrency.lockutils [None req-6f20740a-b9db-46d3-aa96-f2bffb1213de tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.493s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.474421] env[61006]: DEBUG oslo_vmware.api [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337435, 'name': ReconfigVM_Task, 'duration_secs': 0.670827} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.474935] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.475174] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfigured VM to attach interface {{(pid=61006) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 983.545577] env[61006]: DEBUG oslo_vmware.api [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337436, 'name': PowerOnVM_Task, 'duration_secs': 0.553698} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.545857] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.633743] env[61006]: DEBUG oslo_vmware.api [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337437, 'name': PowerOnVM_Task, 'duration_secs': 0.511152} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.634037] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.634231] env[61006]: DEBUG nova.compute.manager [None req-e9a10a9e-b01c-4d89-b174-36a1e87f2284 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.634994] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122a2fa0-db83-4245-a7e5-c6da2c2958a2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.644682] env[61006]: DEBUG nova.compute.manager [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.645541] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.645773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.646008] env[61006]: DEBUG nova.objects.instance [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lazy-loading 'resources' on Instance uuid 3a3aa93a-818e-41d0-896e-72ba97b8b377 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.647393] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48758907-160d-4e52-b89a-2dfdac5b8204 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.700679] env[61006]: DEBUG oslo_concurrency.lockutils [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.700963] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Received event network-vif-deleted-52802f58-a9a0-4d6b-a555-cc9e441ea74b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.701170] env[61006]: INFO nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Neutron deleted interface 52802f58-a9a0-4d6b-a555-cc9e441ea74b; detaching it from the instance and deleting it from the info cache [ 983.701524] env[61006]: DEBUG nova.network.neutron [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.758118] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.758466] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337439, 'name': Rename_Task, 'duration_secs': 0.180601} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.758694] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-715e3d78-b1cd-4e9d-97b8-e89f7c233e3b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.760400] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.760635] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cf223a7-8f8a-4e40-85cd-5ca2408ccef3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.769105] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 983.769105] env[61006]: value = "task-1337440" [ 983.769105] env[61006]: _type = "Task" [ 983.769105] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.769369] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 983.769369] env[61006]: value = "task-1337441" [ 983.769369] env[61006]: _type = "Task" [ 983.769369] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.780493] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337441, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.783603] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.860092] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.860553] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.860837] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "6de1f827-1a19-457b-8600-6546593e55ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.861101] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.861341] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.863926] env[61006]: INFO nova.compute.manager [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Terminating instance [ 983.866134] env[61006]: DEBUG nova.compute.manager [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 983.866367] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.867319] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349e6649-1a49-4499-9e82-74fa1a90b073 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.875190] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.875447] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91bfbf08-da5b-456d-9af0-177ac7a75135 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.881839] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 983.881839] env[61006]: value = "task-1337442" [ 983.881839] env[61006]: _type = "Task" [ 983.881839] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.890636] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337442, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.979941] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e09414a9-9539-4680-8ffb-25fb968d4855 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.905s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.163307] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2a28bebd-05b5-47fb-89e8-51c2f586097d tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 41.635s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.207712] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4be9937-f23f-4a00-a19a-63b9103e2efa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.217262] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f404ff3-20fb-4c4d-b6b0-02624153e106 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.252259] env[61006]: DEBUG nova.compute.manager [req-7d47f474-6feb-45a3-a10b-a15559a47fa2 req-4bf60465-edeb-4433-8907-ee47576d6954 service nova] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Detach interface failed, port_id=52802f58-a9a0-4d6b-a555-cc9e441ea74b, reason: Instance 3a3aa93a-818e-41d0-896e-72ba97b8b377 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 984.287775] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337440, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.288010] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337441, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.292296] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.292572] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.388623] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cc16ef-14aa-480c-bf42-f6e9ff78cfc0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.394181] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337442, 'name': PowerOffVM_Task, 'duration_secs': 0.278168} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.394803] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.395017] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.395291] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47e37e64-3218-4685-9085-6a52dfa357cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.399629] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9721d9-6af6-489f-bb43-893823a7e9c7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.431397] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87f3a13-bb9b-4a14-9950-c942ff5b6471 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.439448] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd959b8c-4816-45ef-8fa3-d30ab69a15e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.453769] env[61006]: DEBUG nova.compute.provider_tree [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.473176] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.473387] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.473577] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleting the datastore file [datastore2] 6de1f827-1a19-457b-8600-6546593e55ca {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.473897] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ebeeaada-bef9-4f30-b1c1-119cfaf6c71b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.480528] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 984.480528] env[61006]: value = "task-1337444" [ 984.480528] env[61006]: _type = "Task" [ 984.480528] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.489674] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.784937] env[61006]: DEBUG oslo_vmware.api [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337440, 'name': PowerOnVM_Task, 'duration_secs': 0.579702} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.788863] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.789100] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cf42ec-8fbf-4918-acc0-08dda93bbfa0 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance 'd027fd07-c409-46e5-857f-cdd2c0479f53' progress to 100 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 984.793399] env[61006]: DEBUG oslo_vmware.api [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337441, 'name': PowerOnVM_Task, 'duration_secs': 0.729785} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.794904] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.795636] env[61006]: INFO nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Took 7.39 seconds to spawn the instance on the hypervisor. [ 984.795636] env[61006]: DEBUG nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.796219] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a975398-052b-4729-8912-23b5b6a04b33 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.798829] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.957264] env[61006]: DEBUG nova.scheduler.client.report [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.990762] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.003393] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.003762] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.004036] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.004216] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.004435] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.006862] env[61006]: INFO nova.compute.manager [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Terminating instance [ 985.008942] env[61006]: DEBUG nova.compute.manager [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.009170] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.010125] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b485b78-f34c-4721-b825-d6dec288d070 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.017868] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.018118] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a19280d-43f3-4307-948d-7d8403901a8e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.025730] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 985.025730] env[61006]: value = "task-1337445" [ 985.025730] env[61006]: _type = "Task" [ 985.025730] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.033338] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.298219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.298457] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.314486] env[61006]: INFO nova.compute.manager [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Took 13.87 seconds to build instance. [ 985.321188] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.462116] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.464622] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.144s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.466169] env[61006]: INFO nova.compute.claims [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.480747] env[61006]: INFO nova.scheduler.client.report [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Deleted allocations for instance 3a3aa93a-818e-41d0-896e-72ba97b8b377 [ 985.492592] env[61006]: DEBUG oslo_vmware.api [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.560604} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.492835] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.493447] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.493447] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.493447] env[61006]: INFO nova.compute.manager [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Took 1.63 seconds to destroy the instance on the hypervisor. [ 985.493640] env[61006]: DEBUG oslo.service.loopingcall [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.493800] env[61006]: DEBUG nova.compute.manager [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.495199] env[61006]: DEBUG nova.network.neutron [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.536899] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337445, 'name': PowerOffVM_Task, 'duration_secs': 0.392086} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.537185] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.537361] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.537640] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4ba5720-e789-452d-84ca-4a6af39e8ad1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.601160] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.601487] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.601607] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleting the datastore file [datastore2] 524f3fd1-1e71-40c0-96c2-0acac5055e01 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.601889] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56b3b0bb-a8b1-4d45-affe-e20273006959 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.608485] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 985.608485] env[61006]: value = "task-1337447" [ 985.608485] env[61006]: _type = "Task" [ 985.608485] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.617277] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.790624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "095f85df-7265-46a9-bdbb-963292384493" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.805047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.805047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.805687] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc4cd99-234d-420f-b4fc-7e36a0eecbf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.823122] env[61006]: DEBUG oslo_concurrency.lockutils [None req-24ec43d3-6f93-435f-a228-8ee601006b00 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.390s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.824744] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.034s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.824744] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "095f85df-7265-46a9-bdbb-963292384493-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.824744] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.824969] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.827167] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e1039f-5571-445e-bac5-ceaaa5ab0cfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.830106] env[61006]: INFO nova.compute.manager [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Terminating instance [ 985.833661] env[61006]: DEBUG nova.compute.manager [req-75d4b732-ac22-4602-9b4d-fb79f17a3c23 req-182010a4-8ea7-4739-9976-01c69b8bd463 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Received event network-vif-deleted-de9a3378-0e2c-485b-842b-d44c829bee0f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.833897] env[61006]: INFO nova.compute.manager [req-75d4b732-ac22-4602-9b4d-fb79f17a3c23 req-182010a4-8ea7-4739-9976-01c69b8bd463 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Neutron deleted interface de9a3378-0e2c-485b-842b-d44c829bee0f; detaching it from the instance and deleting it from the info cache [ 985.835033] env[61006]: DEBUG nova.network.neutron [req-75d4b732-ac22-4602-9b4d-fb79f17a3c23 req-182010a4-8ea7-4739-9976-01c69b8bd463 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.835404] env[61006]: DEBUG nova.compute.manager [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.835658] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.837257] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fffbbc-ad30-4f98-962f-8eeb90ca2ac0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.862588] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfiguring VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 985.863861] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5901de8b-08a4-408b-a145-adebc80be809 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.879605] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.880258] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c901f937-242e-424d-abe1-18c2e28d7e57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.885445] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 985.885445] env[61006]: value = "task-1337448" [ 985.885445] env[61006]: _type = "Task" [ 985.885445] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.886615] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 985.886615] env[61006]: value = "task-1337449" [ 985.886615] env[61006]: _type = "Task" [ 985.886615] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.896844] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.899635] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.991283] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d6c030bc-b052-4a71-a9b5-be4e0e5de74f tempest-ServerMetadataNegativeTestJSON-912044780 tempest-ServerMetadataNegativeTestJSON-912044780-project-member] Lock "3a3aa93a-818e-41d0-896e-72ba97b8b377" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.375s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.118142] env[61006]: DEBUG oslo_vmware.api [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164047} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.118237] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.118426] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.118610] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.118783] env[61006]: INFO nova.compute.manager [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 1.11 seconds to destroy the instance on the hypervisor. [ 986.119035] env[61006]: DEBUG oslo.service.loopingcall [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.119236] env[61006]: DEBUG nova.compute.manager [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.119328] env[61006]: DEBUG nova.network.neutron [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.301032] env[61006]: DEBUG nova.network.neutron [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.341239] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34c9b8a3-ebdb-4c46-9ef7-6c9de5950c92 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.352604] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff014ef-e482-4bfb-a687-44a40a3be53a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.392177] env[61006]: DEBUG nova.compute.manager [req-75d4b732-ac22-4602-9b4d-fb79f17a3c23 req-182010a4-8ea7-4739-9976-01c69b8bd463 service nova] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Detach interface failed, port_id=de9a3378-0e2c-485b-842b-d44c829bee0f, reason: Instance 6de1f827-1a19-457b-8600-6546593e55ca could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.405709] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.411844] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.713796] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1d466e-01bc-46b3-b54a-b147936b2dc2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.722786] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa20c00-314b-410a-983a-762828489c5c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.757259] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57cc2747-a9c2-4278-b9bf-bc3fcd223438 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.765901] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef1eb7c-1424-4ac6-b76f-bac4164edad5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.779830] env[61006]: DEBUG nova.compute.provider_tree [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.803549] env[61006]: INFO nova.compute.manager [-] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Took 1.31 seconds to deallocate network for instance. [ 986.904725] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337449, 'name': PowerOffVM_Task, 'duration_secs': 0.573125} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.907971] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.908216] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.908531] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.908758] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a9a6b8e-0812-45cc-9e88-011d4a58f683 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.982908] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.982908] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.982908] env[61006]: DEBUG nova.compute.manager [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Going to confirm migration 4 {{(pid=61006) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 986.987903] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.987903] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.988050] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore2] 095f85df-7265-46a9-bdbb-963292384493 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.988226] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53a6dc83-d7f7-48f0-91f4-00edfa6a27c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.996085] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 986.996085] env[61006]: value = "task-1337451" [ 986.996085] env[61006]: _type = "Task" [ 986.996085] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.005867] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337451, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.286292] env[61006]: DEBUG nova.scheduler.client.report [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.310072] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.358906] env[61006]: DEBUG nova.network.neutron [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.401525] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.507033] env[61006]: DEBUG oslo_vmware.api [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337451, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172881} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.507348] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.507573] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.508173] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.508173] env[61006]: INFO nova.compute.manager [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 095f85df-7265-46a9-bdbb-963292384493] Took 1.67 seconds to destroy the instance on the hypervisor. [ 987.508263] env[61006]: DEBUG oslo.service.loopingcall [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.508570] env[61006]: DEBUG nova.compute.manager [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.508624] env[61006]: DEBUG nova.network.neutron [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.554221] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.554575] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquired lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.554774] env[61006]: DEBUG nova.network.neutron [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.554966] env[61006]: DEBUG nova.objects.instance [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lazy-loading 'info_cache' on Instance uuid d027fd07-c409-46e5-857f-cdd2c0479f53 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.789953] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.790601] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.794551] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.485s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.794819] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.826111] env[61006]: INFO nova.scheduler.client.report [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocations for instance 6de1f827-1a19-457b-8600-6546593e55ca [ 987.834562] env[61006]: DEBUG nova.compute.manager [req-921ec704-d698-4e62-a285-49d8f11e2dd3 req-523a3b28-cfd8-4639-bea6-c399471b050e service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Received event network-vif-deleted-45c0cda2-36f4-4b9a-9e88-07b136107358 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.834827] env[61006]: INFO nova.compute.manager [req-921ec704-d698-4e62-a285-49d8f11e2dd3 req-523a3b28-cfd8-4639-bea6-c399471b050e service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Neutron deleted interface 45c0cda2-36f4-4b9a-9e88-07b136107358; detaching it from the instance and deleting it from the info cache [ 987.835152] env[61006]: DEBUG nova.network.neutron [req-921ec704-d698-4e62-a285-49d8f11e2dd3 req-523a3b28-cfd8-4639-bea6-c399471b050e service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.525069] env[61006]: INFO nova.compute.manager [-] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Took 2.41 seconds to deallocate network for instance. [ 988.531027] env[61006]: DEBUG nova.compute.utils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.531027] env[61006]: DEBUG nova.network.neutron [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.532618] env[61006]: DEBUG nova.compute.manager [req-b153d4c1-824a-4ad3-bee9-54058eaf44fc req-e5d7e6f2-82ac-4871-9643-135ad0a593f9 service nova] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Received event network-vif-deleted-3d9ba3b1-a27e-4002-bd16-cfd881fca452 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.533424] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a6bf8e03-490a-476b-abd9-7e2d9682e387 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "6de1f827-1a19-457b-8600-6546593e55ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.673s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.538607] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.538794] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 988.540853] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74a1741c-1d72-4070-a87f-e33f6f6807ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.552630] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.556270] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3e1a15-e570-4813-87ef-2c647a225cae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.584237] env[61006]: DEBUG nova.compute.manager [req-921ec704-d698-4e62-a285-49d8f11e2dd3 req-523a3b28-cfd8-4639-bea6-c399471b050e service nova] [instance: 095f85df-7265-46a9-bdbb-963292384493] Detach interface failed, port_id=45c0cda2-36f4-4b9a-9e88-07b136107358, reason: Instance 095f85df-7265-46a9-bdbb-963292384493 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 988.587337] env[61006]: DEBUG nova.policy [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cca6e2806aa45208ae618f6a78ccc0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe9eabeec9a941e68a9eae559e24ff4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.844342] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Successfully created port: 368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.035624] env[61006]: INFO nova.compute.manager [-] [instance: 095f85df-7265-46a9-bdbb-963292384493] Took 1.53 seconds to deallocate network for instance. [ 989.036137] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.040417] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.052860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.052860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.052860] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.083529] env[61006]: INFO nova.scheduler.client.report [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted allocations for instance 524f3fd1-1e71-40c0-96c2-0acac5055e01 [ 989.284928] env[61006]: DEBUG nova.network.neutron [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [{"id": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "address": "fa:16:3e:d4:1c:92", "network": {"id": "305351dc-89cf-42db-a0cf-dc1c8062fbf5", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-2127489654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8afbe151c21743bfb40dc12ba384db28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37333dc2-982e-45e9-9dda-0c18417d7fa6", "external-id": "nsx-vlan-transportzone-227", "segmentation_id": 227, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a7b1ed1-7d", "ovs_interfaceid": "1a7b1ed1-7ded-4bd0-880f-96439cf27513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.537165] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.547633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.547882] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.548108] env[61006]: DEBUG nova.objects.instance [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid 095f85df-7265-46a9-bdbb-963292384493 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.594329] env[61006]: DEBUG oslo_concurrency.lockutils [None req-b9012581-f6e8-4c00-aa50-335fc4b4b0d7 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "524f3fd1-1e71-40c0-96c2-0acac5055e01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.590s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.788024] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Releasing lock "refresh_cache-d027fd07-c409-46e5-857f-cdd2c0479f53" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.788024] env[61006]: DEBUG nova.objects.instance [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lazy-loading 'migration_context' on Instance uuid d027fd07-c409-46e5-857f-cdd2c0479f53 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.036416] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.050791] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.081595] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.081890] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.082060] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.082254] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.082438] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.082596] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.082806] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.082963] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.083143] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.083305] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.083478] env[61006]: DEBUG nova.virt.hardware [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.084368] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad3f06a-755d-4a37-9938-670c3aad0d1b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.092749] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa22c3b-9665-497a-833b-712913fc0f9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.263167] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af90790-08b8-424b-ace9-789ed3a093b6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.271128] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cae55e-0bde-4f8f-bee6-d39afe7bd6f1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.301844] env[61006]: DEBUG nova.objects.base [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 990.302578] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14863896-ad8d-43ea-9faa-e4865d654c20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.306180] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0707f41-6b82-4e9c-933b-9848e729b5c4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.328299] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99bc6f48-8a31-4097-a7b4-3586e959174c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.333614] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbeaaa9-26a2-4211-ba70-f29d256802d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.340349] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 990.340349] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528b9132-7590-da9b-7901-5c4df9faf4fa" [ 990.340349] env[61006]: _type = "Task" [ 990.340349] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.347128] env[61006]: DEBUG nova.compute.provider_tree [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.356916] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528b9132-7590-da9b-7901-5c4df9faf4fa, 'name': SearchDatastore_Task, 'duration_secs': 0.007409} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.357320] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.366717] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Successfully updated port: 368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.537120] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.597937] env[61006]: DEBUG nova.compute.manager [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Received event network-vif-plugged-368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.598221] env[61006]: DEBUG oslo_concurrency.lockutils [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] Acquiring lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.598390] env[61006]: DEBUG oslo_concurrency.lockutils [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.598557] env[61006]: DEBUG oslo_concurrency.lockutils [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.598734] env[61006]: DEBUG nova.compute.manager [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] No waiting events found dispatching network-vif-plugged-368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.598896] env[61006]: WARNING nova.compute.manager [req-0f8d920c-a37c-449f-8315-edba3ac7fc74 req-08878d5e-0796-4e13-a7e8-7d2c039ffc70 service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Received unexpected event network-vif-plugged-368436a2-9db8-429d-9f98-d4f5baf42930 for instance with vm_state building and task_state spawning. [ 990.850120] env[61006]: DEBUG nova.scheduler.client.report [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.870291] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.870507] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.870665] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.038827] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.092032] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.092203] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.354619] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.357048] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.382551] env[61006]: INFO nova.scheduler.client.report [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance 095f85df-7265-46a9-bdbb-963292384493 [ 991.404762] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.415447] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.415658] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.539027] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.540869] env[61006]: DEBUG nova.network.neutron [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Updating instance_info_cache with network_info: [{"id": "368436a2-9db8-429d-9f98-d4f5baf42930", "address": "fa:16:3e:0b:2b:ca", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap368436a2-9d", "ovs_interfaceid": "368436a2-9db8-429d-9f98-d4f5baf42930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.594886] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 991.891366] env[61006]: DEBUG oslo_concurrency.lockutils [None req-de7233f9-8cfb-4813-ac53-0277834ba108 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "095f85df-7265-46a9-bdbb-963292384493" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.067s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.918690] env[61006]: DEBUG nova.compute.utils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.015546] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260d570a-6bf0-43c7-8e33-aa586b663435 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.022969] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5651b68f-265a-4ce7-b028-48eb38150bb6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.052602] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.052901] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Instance network_info: |[{"id": "368436a2-9db8-429d-9f98-d4f5baf42930", "address": "fa:16:3e:0b:2b:ca", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap368436a2-9d", "ovs_interfaceid": "368436a2-9db8-429d-9f98-d4f5baf42930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.056378] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:2b:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb68953b-dee5-4d9d-b47b-277336ba76dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '368436a2-9db8-429d-9f98-d4f5baf42930', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.063665] env[61006]: DEBUG oslo.service.loopingcall [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.064336] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664e6732-b687-43ed-9041-6c6fe3ce0fc0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.066913] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.067870] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7310070-fbd5-425c-8ecf-63351d3b72c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.087123] env[61006]: DEBUG oslo_vmware.api [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337448, 'name': ReconfigVM_Task, 'duration_secs': 5.751823} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.089643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.089915] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Reconfigured VM to detach interface {{(pid=61006) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 992.093099] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf903738-01cf-42eb-8962-203bc6de1e22 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.097873] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.097873] env[61006]: value = "task-1337452" [ 992.097873] env[61006]: _type = "Task" [ 992.097873] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.113172] env[61006]: DEBUG nova.compute.provider_tree [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.121016] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337452, 'name': CreateVM_Task} progress is 10%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.125263] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.421943] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.608716] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337452, 'name': CreateVM_Task} progress is 25%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.617287] env[61006]: DEBUG nova.scheduler.client.report [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.109325] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337452, 'name': CreateVM_Task, 'duration_secs': 0.651934} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.109604] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.110261] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.110475] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.110797] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.111693] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f602dac0-7e4f-43a4-91a3-6e2e36405c3e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.116674] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 993.116674] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5204b5b2-4448-7c4d-3a26-49cbccfa795c" [ 993.116674] env[61006]: _type = "Task" [ 993.116674] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.127753] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5204b5b2-4448-7c4d-3a26-49cbccfa795c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.175716] env[61006]: DEBUG nova.compute.manager [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Received event network-changed-368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.175891] env[61006]: DEBUG nova.compute.manager [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Refreshing instance network info cache due to event network-changed-368436a2-9db8-429d-9f98-d4f5baf42930. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.176124] env[61006]: DEBUG oslo_concurrency.lockutils [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] Acquiring lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.176270] env[61006]: DEBUG oslo_concurrency.lockutils [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] Acquired lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.176433] env[61006]: DEBUG nova.network.neutron [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Refreshing network info cache for port 368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.400177] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.400382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquired lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.400564] env[61006]: DEBUG nova.network.neutron [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.484026] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.484185] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.484436] env[61006]: INFO nova.compute.manager [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Attaching volume 5501507e-d513-49f9-9e05-434386d674ca to /dev/sdb [ 993.514839] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09044456-a74c-493f-99ad-bc50d238e44f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.521815] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657c1518-1bcf-4461-9e62-65aebec41cb0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.535177] env[61006]: DEBUG nova.virt.block_device [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating existing volume attachment record: 18645d44-4ab8-48b5-8a05-5a36c01af9d2 {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 993.626816] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5204b5b2-4448-7c4d-3a26-49cbccfa795c, 'name': SearchDatastore_Task, 'duration_secs': 0.013298} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.627143] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.627381] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.627621] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.627773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.627958] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.629115] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.272s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.631744] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97e9e94d-7433-4634-ae42-a99b8a548401 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.633751] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.509s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.635170] env[61006]: INFO nova.compute.claims [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.644017] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.644214] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.644923] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f35948f3-a7ed-4bc4-9f5b-7967871051c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.650072] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 993.650072] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a7238f-6c12-9dd1-a6ca-25f340cc3545" [ 993.650072] env[61006]: _type = "Task" [ 993.650072] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.658217] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a7238f-6c12-9dd1-a6ca-25f340cc3545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.729550] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.729840] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.730067] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.730277] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.731528] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.732792] env[61006]: INFO nova.compute.manager [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Terminating instance [ 993.735053] env[61006]: DEBUG nova.compute.manager [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.735053] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.735718] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1e36ce-2473-4e0b-89ae-3c8c9df56bff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.743162] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.743443] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc49e7aa-de8c-42c6-bddc-ee019850662f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.749701] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 993.749701] env[61006]: value = "task-1337454" [ 993.749701] env[61006]: _type = "Task" [ 993.749701] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.757782] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.787930] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.788654] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.788654] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.788654] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.789038] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.791515] env[61006]: INFO nova.compute.manager [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Terminating instance [ 993.793730] env[61006]: DEBUG nova.compute.manager [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.793730] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.793983] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9294917c-aa7a-485f-a11b-ef3546182107 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.800911] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.801209] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9884ba1-d44c-4f03-99cd-7b410171d781 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.808030] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 993.808030] env[61006]: value = "task-1337455" [ 993.808030] env[61006]: _type = "Task" [ 993.808030] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.817048] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337455, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.906302] env[61006]: DEBUG nova.network.neutron [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Updated VIF entry in instance network info cache for port 368436a2-9db8-429d-9f98-d4f5baf42930. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 993.906663] env[61006]: DEBUG nova.network.neutron [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Updating instance_info_cache with network_info: [{"id": "368436a2-9db8-429d-9f98-d4f5baf42930", "address": "fa:16:3e:0b:2b:ca", "network": {"id": "39ba5bce-e81d-44b9-9e64-12715d558fca", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1036765984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe9eabeec9a941e68a9eae559e24ff4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap368436a2-9d", "ovs_interfaceid": "368436a2-9db8-429d-9f98-d4f5baf42930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.119856] env[61006]: INFO nova.network.neutron [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Port f07eb7f2-10e5-4957-a5c8-3792f6d15b26 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 994.120305] env[61006]: DEBUG nova.network.neutron [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [{"id": "5d945d09-25a9-4087-96d5-f0be90bcddff", "address": "fa:16:3e:7d:f3:fa", "network": {"id": "98dcba29-9f32-40eb-b526-46e091c2a330", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1847008146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "85b03e26e0034e30b74761724d0a39e3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d945d09-25", "ovs_interfaceid": "5d945d09-25a9-4087-96d5-f0be90bcddff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.163976] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a7238f-6c12-9dd1-a6ca-25f340cc3545, 'name': SearchDatastore_Task, 'duration_secs': 0.009287} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.164476] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17ea8223-d966-45e7-a11c-e98bbb9b85c4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.169160] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 994.169160] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e945d0-e542-733d-1c8a-fdb3a1099beb" [ 994.169160] env[61006]: _type = "Task" [ 994.169160] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.176726] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e945d0-e542-733d-1c8a-fdb3a1099beb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.194913] env[61006]: INFO nova.scheduler.client.report [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocation for migration 690dbea6-80cc-4e32-9907-818643d640d4 [ 994.259374] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337454, 'name': PowerOffVM_Task, 'duration_secs': 0.207656} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.259686] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.259857] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.260131] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45a8d4eb-cd12-4bf5-82f6-a8b5e258b987 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.318600] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337455, 'name': PowerOffVM_Task, 'duration_secs': 0.192838} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.319685] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.319868] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.320145] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.320328] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.320534] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleting the datastore file [datastore2] f048bd9c-048b-4439-bbbf-dfcee7f18d84 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.320765] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c308c39-7825-4495-9ed9-e53eac0aeafc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.322142] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5016f7b7-b4da-410a-bdba-996e988681ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.328325] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 994.328325] env[61006]: value = "task-1337460" [ 994.328325] env[61006]: _type = "Task" [ 994.328325] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.335741] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337460, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.380104] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.380366] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.380665] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore1] 0b2b6d52-8340-4f85-9fc0-ab1f2b079031 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.380952] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3e69064-3d95-470e-b66e-72b47fb80777 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.387271] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 994.387271] env[61006]: value = "task-1337461" [ 994.387271] env[61006]: _type = "Task" [ 994.387271] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.395181] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.408921] env[61006]: DEBUG oslo_concurrency.lockutils [req-d23783cd-4fdd-439e-bee8-9e4d0f71d4c1 req-960e067b-805a-4002-86e0-e6170b682c5c service nova] Releasing lock "refresh_cache-feb3c68c-4c03-40cd-89a5-3310bfcab359" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.623054] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Releasing lock "refresh_cache-f048bd9c-048b-4439-bbbf-dfcee7f18d84" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.678835] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e945d0-e542-733d-1c8a-fdb3a1099beb, 'name': SearchDatastore_Task, 'duration_secs': 0.009089} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.679119] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.679373] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] feb3c68c-4c03-40cd-89a5-3310bfcab359/feb3c68c-4c03-40cd-89a5-3310bfcab359.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.681580] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1d7149b-dad0-4d38-84c1-beb14e5f1862 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.688423] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 994.688423] env[61006]: value = "task-1337462" [ 994.688423] env[61006]: _type = "Task" [ 994.688423] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.698331] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337462, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.700273] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.718s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.782686] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cc72b1-8cb6-4520-92b4-b9404b7c5560 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.790343] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cbce6c-a94c-40bb-b5b6-4db4cbbbca16 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.822348] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f058e9d-56fa-4fb6-aa97-f10529031807 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.836280] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6143033-6805-495f-8d50-d2e409fb5657 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.855952] env[61006]: DEBUG nova.compute.provider_tree [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.864804] env[61006]: DEBUG oslo_vmware.api [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337460, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153779} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.866335] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.866706] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 994.867019] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 994.867269] env[61006]: INFO nova.compute.manager [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Took 1.13 seconds to destroy the instance on the hypervisor. [ 994.867564] env[61006]: DEBUG oslo.service.loopingcall [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.868113] env[61006]: DEBUG nova.compute.manager [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.868230] env[61006]: DEBUG nova.network.neutron [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 994.897628] env[61006]: DEBUG oslo_vmware.api [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133582} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.897969] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.898200] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 994.898389] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 994.898596] env[61006]: INFO nova.compute.manager [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Took 1.11 seconds to destroy the instance on the hypervisor. [ 994.899857] env[61006]: DEBUG oslo.service.loopingcall [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.900176] env[61006]: DEBUG nova.compute.manager [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.900281] env[61006]: DEBUG nova.network.neutron [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.127138] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce2bd2df-0dbf-4f7f-b896-f929f8883b4c tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "interface-f048bd9c-048b-4439-bbbf-dfcee7f18d84-f07eb7f2-10e5-4957-a5c8-3792f6d15b26" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.828s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.198032] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337462, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48225} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.198321] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore1] feb3c68c-4c03-40cd-89a5-3310bfcab359/feb3c68c-4c03-40cd-89a5-3310bfcab359.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.198541] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.198803] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4b57f73-066a-4236-a7ad-55e05c83251c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.206656] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 995.206656] env[61006]: value = "task-1337463" [ 995.206656] env[61006]: _type = "Task" [ 995.206656] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.208056] env[61006]: DEBUG nova.compute.manager [req-5737774f-7445-4124-95c6-882b93786697 req-cfd12695-d1cb-44c1-9c5f-3aa0da7d5bd4 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Received event network-vif-deleted-bfba9043-9516-43cb-9874-048f6b5076b7 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.208275] env[61006]: INFO nova.compute.manager [req-5737774f-7445-4124-95c6-882b93786697 req-cfd12695-d1cb-44c1-9c5f-3aa0da7d5bd4 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Neutron deleted interface bfba9043-9516-43cb-9874-048f6b5076b7; detaching it from the instance and deleting it from the info cache [ 995.208478] env[61006]: DEBUG nova.network.neutron [req-5737774f-7445-4124-95c6-882b93786697 req-cfd12695-d1cb-44c1-9c5f-3aa0da7d5bd4 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.220972] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337463, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.359478] env[61006]: DEBUG nova.scheduler.client.report [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.649595] env[61006]: DEBUG nova.compute.manager [req-abefe862-ca9d-4973-8391-172e411291f0 req-e6de17a7-70c1-4a40-9468-4ce6aa2407ca service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Received event network-vif-deleted-5d945d09-25a9-4087-96d5-f0be90bcddff {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.649595] env[61006]: INFO nova.compute.manager [req-abefe862-ca9d-4973-8391-172e411291f0 req-e6de17a7-70c1-4a40-9468-4ce6aa2407ca service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Neutron deleted interface 5d945d09-25a9-4087-96d5-f0be90bcddff; detaching it from the instance and deleting it from the info cache [ 995.649595] env[61006]: DEBUG nova.network.neutron [req-abefe862-ca9d-4973-8391-172e411291f0 req-e6de17a7-70c1-4a40-9468-4ce6aa2407ca service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.677923] env[61006]: DEBUG nova.network.neutron [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.711149] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23187023-9df2-4f8b-ab98-19536ae334ce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.722821] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337463, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.23471} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.723155] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.725905] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c400c6f8-8547-4fc0-91cd-596dfad2b661 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.736858] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bf0241-838d-4350-9614-c902c4905104 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.758720] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] feb3c68c-4c03-40cd-89a5-3310bfcab359/feb3c68c-4c03-40cd-89a5-3310bfcab359.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.767450] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0a6641c-960b-4cd7-ae10-4c35b4dbcd79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.781913] env[61006]: DEBUG nova.compute.manager [req-5737774f-7445-4124-95c6-882b93786697 req-cfd12695-d1cb-44c1-9c5f-3aa0da7d5bd4 service nova] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Detach interface failed, port_id=bfba9043-9516-43cb-9874-048f6b5076b7, reason: Instance 0b2b6d52-8340-4f85-9fc0-ab1f2b079031 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 995.787351] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 995.787351] env[61006]: value = "task-1337464" [ 995.787351] env[61006]: _type = "Task" [ 995.787351] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.795113] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.864742] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.865337] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.010928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.011178] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.011394] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.011585] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.011761] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.014200] env[61006]: INFO nova.compute.manager [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Terminating instance [ 996.016289] env[61006]: DEBUG nova.compute.manager [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.016486] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.017340] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6a91d6-2c5e-4158-8adf-6709c61fb33b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.024866] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.025111] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74c8e395-a09a-46c9-ae6c-45d99f7bdc48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.030620] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 996.030620] env[61006]: value = "task-1337466" [ 996.030620] env[61006]: _type = "Task" [ 996.030620] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.039427] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.106752] env[61006]: DEBUG nova.network.neutron [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.152030] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dd472ed-3720-4eda-9868-2c8920fc6a58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.160058] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13576594-c3cf-4fb9-b6bd-b0d582e0f553 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.180107] env[61006]: INFO nova.compute.manager [-] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Took 1.28 seconds to deallocate network for instance. [ 996.191102] env[61006]: DEBUG nova.compute.manager [req-abefe862-ca9d-4973-8391-172e411291f0 req-e6de17a7-70c1-4a40-9468-4ce6aa2407ca service nova] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Detach interface failed, port_id=5d945d09-25a9-4087-96d5-f0be90bcddff, reason: Instance f048bd9c-048b-4439-bbbf-dfcee7f18d84 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.297338] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337464, 'name': ReconfigVM_Task, 'duration_secs': 0.342316} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.297338] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Reconfigured VM instance instance-00000064 to attach disk [datastore1] feb3c68c-4c03-40cd-89a5-3310bfcab359/feb3c68c-4c03-40cd-89a5-3310bfcab359.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.297788] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8700ad26-ce27-40c3-bbf2-31c9602d6b89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.303678] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 996.303678] env[61006]: value = "task-1337467" [ 996.303678] env[61006]: _type = "Task" [ 996.303678] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.310872] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337467, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.370361] env[61006]: DEBUG nova.compute.utils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.371829] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.372009] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.419260] env[61006]: DEBUG nova.policy [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7707895279cb4c6ebc07bf3570f3f881', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c673bf1b8a437fbfbfd34e912a8f37', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.540703] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337466, 'name': PowerOffVM_Task, 'duration_secs': 0.216234} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.541052] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.541317] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.541492] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-694bfdb1-abaa-4270-8bf9-d6d1c6d1072f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.608286] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.608565] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.608805] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleting the datastore file [datastore2] d027fd07-c409-46e5-857f-cdd2c0479f53 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.609313] env[61006]: INFO nova.compute.manager [-] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Took 1.74 seconds to deallocate network for instance. [ 996.609547] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2358bf1-73f9-47e6-ba72-787094c57f07 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.619072] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for the task: (returnval){ [ 996.619072] env[61006]: value = "task-1337469" [ 996.619072] env[61006]: _type = "Task" [ 996.619072] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.628386] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.695976] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.696398] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.696524] env[61006]: DEBUG nova.objects.instance [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid 0b2b6d52-8340-4f85-9fc0-ab1f2b079031 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.712430] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Successfully created port: cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.813260] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337467, 'name': Rename_Task, 'duration_secs': 0.13784} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.813508] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 996.813763] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27c73cd5-1434-495d-911f-5f3e48da7c97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.819500] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 996.819500] env[61006]: value = "task-1337470" [ 996.819500] env[61006]: _type = "Task" [ 996.819500] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.826294] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337470, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.875543] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.122642] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.131490] env[61006]: DEBUG oslo_vmware.api [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Task: {'id': task-1337469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164029} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.131778] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.131967] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.132164] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.132347] env[61006]: INFO nova.compute.manager [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Took 1.12 seconds to destroy the instance on the hypervisor. [ 997.132645] env[61006]: DEBUG oslo.service.loopingcall [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.132848] env[61006]: DEBUG nova.compute.manager [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.132942] env[61006]: DEBUG nova.network.neutron [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.333245] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337470, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.359229] env[61006]: DEBUG nova.compute.manager [req-f7802340-79a1-48bb-b97b-aa248303708e req-39bc2730-4ad6-4d33-82f2-ef9f0a3a8a96 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Received event network-vif-deleted-1a7b1ed1-7ded-4bd0-880f-96439cf27513 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.359428] env[61006]: INFO nova.compute.manager [req-f7802340-79a1-48bb-b97b-aa248303708e req-39bc2730-4ad6-4d33-82f2-ef9f0a3a8a96 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Neutron deleted interface 1a7b1ed1-7ded-4bd0-880f-96439cf27513; detaching it from the instance and deleting it from the info cache [ 997.359625] env[61006]: DEBUG nova.network.neutron [req-f7802340-79a1-48bb-b97b-aa248303708e req-39bc2730-4ad6-4d33-82f2-ef9f0a3a8a96 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.362338] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0616ae48-a55f-4bb8-a4bc-70205eb26194 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.369909] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780bebed-64c5-4060-8a34-62e291202975 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.403592] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0124cd-425e-4d0f-b423-2b2be5c7ecd1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.411600] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746b85ca-eb02-4e62-9aab-9b3d101ddada {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.425290] env[61006]: DEBUG nova.compute.provider_tree [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.832942] env[61006]: DEBUG oslo_vmware.api [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337470, 'name': PowerOnVM_Task, 'duration_secs': 0.745057} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.833236] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 997.833464] env[61006]: INFO nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Took 7.78 seconds to spawn the instance on the hypervisor. [ 997.833619] env[61006]: DEBUG nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.834384] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5193ada8-1b3d-4662-882a-779ef3908458 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.839864] env[61006]: DEBUG nova.network.neutron [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.866350] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe8b79ad-48c6-469c-807a-a7a46633d3d4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.875155] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bcc67b-463b-4b80-9467-04ede4561403 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.903340] env[61006]: DEBUG nova.compute.manager [req-f7802340-79a1-48bb-b97b-aa248303708e req-39bc2730-4ad6-4d33-82f2-ef9f0a3a8a96 service nova] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Detach interface failed, port_id=1a7b1ed1-7ded-4bd0-880f-96439cf27513, reason: Instance d027fd07-c409-46e5-857f-cdd2c0479f53 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 997.904622] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 997.929465] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.929721] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.929925] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.930127] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.930277] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.930429] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.930676] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.930849] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.931029] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.931197] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.931370] env[61006]: DEBUG nova.virt.hardware [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.932190] env[61006]: DEBUG nova.scheduler.client.report [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.935706] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a5923f-feed-43c9-ad40-4a292ddf6016 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.944228] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbab5149-e3e9-49c7-8cb3-4a32a08d2678 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.082485] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 998.082749] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285417', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'name': 'volume-5501507e-d513-49f9-9e05-434386d674ca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '02408ca9-c580-444a-9608-a752146ca499', 'attached_at': '', 'detached_at': '', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'serial': '5501507e-d513-49f9-9e05-434386d674ca'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 998.083672] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61372533-aebe-4dd9-9a37-a7f4143bd031 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.103975] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2923f7a-8194-4518-b40d-b7aa30cc0f47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.105720] env[61006]: DEBUG nova.compute.manager [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Received event network-vif-plugged-cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.105862] env[61006]: DEBUG oslo_concurrency.lockutils [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.106086] env[61006]: DEBUG oslo_concurrency.lockutils [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.106260] env[61006]: DEBUG oslo_concurrency.lockutils [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.106425] env[61006]: DEBUG nova.compute.manager [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] No waiting events found dispatching network-vif-plugged-cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 998.106587] env[61006]: WARNING nova.compute.manager [req-f99e45d1-b1a8-48de-a157-a0b42d9a2410 req-69c568de-3a43-4786-bc32-2971a2ed6ba1 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Received unexpected event network-vif-plugged-cdb40136-9153-4611-9dc5-7bde778a2b5a for instance with vm_state building and task_state spawning. [ 998.128972] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] volume-5501507e-d513-49f9-9e05-434386d674ca/volume-5501507e-d513-49f9-9e05-434386d674ca.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.129505] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32c63b6c-7152-4134-aa95-7372640bea41 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.150278] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 998.150278] env[61006]: value = "task-1337471" [ 998.150278] env[61006]: _type = "Task" [ 998.150278] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.158061] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.185986] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Successfully updated port: cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.348686] env[61006]: INFO nova.compute.manager [-] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Took 1.22 seconds to deallocate network for instance. [ 998.354786] env[61006]: INFO nova.compute.manager [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Took 13.05 seconds to build instance. [ 998.439645] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.441972] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.320s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.442227] env[61006]: DEBUG nova.objects.instance [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'resources' on Instance uuid f048bd9c-048b-4439-bbbf-dfcee7f18d84 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.462602] env[61006]: INFO nova.scheduler.client.report [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance 0b2b6d52-8340-4f85-9fc0-ab1f2b079031 [ 998.663370] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337471, 'name': ReconfigVM_Task, 'duration_secs': 0.43999} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.663829] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfigured VM instance instance-0000005e to attach disk [datastore1] volume-5501507e-d513-49f9-9e05-434386d674ca/volume-5501507e-d513-49f9-9e05-434386d674ca.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.671617] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e050398-b4f6-440d-a12e-fd73da1e69a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.689390] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.689390] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.689390] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 998.696095] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 998.696095] env[61006]: value = "task-1337472" [ 998.696095] env[61006]: _type = "Task" [ 998.696095] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.709410] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337472, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.857043] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.857945] env[61006]: DEBUG oslo_concurrency.lockutils [None req-34397c1d-673a-46f8-bbfd-f04779cb6a10 tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.565s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.972034] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c102ac25-7e50-4b08-a986-501689109278 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "0b2b6d52-8340-4f85-9fc0-ab1f2b079031" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.184s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.101378] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080e203f-02cb-4b78-a308-a21a8e9ad4a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.110021] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e8b3fa-ecff-4912-b4e6-7dffb68da2ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.142111] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446c0a49-9dd0-4316-8509-f1a8dd24e5f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.149454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ac9f98-f961-4427-8dad-e4717007d784 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.163249] env[61006]: DEBUG nova.compute.provider_tree [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.208520] env[61006]: DEBUG oslo_vmware.api [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337472, 'name': ReconfigVM_Task, 'duration_secs': 0.141602} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.208932] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285417', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'name': 'volume-5501507e-d513-49f9-9e05-434386d674ca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '02408ca9-c580-444a-9608-a752146ca499', 'attached_at': '', 'detached_at': '', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'serial': '5501507e-d513-49f9-9e05-434386d674ca'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 999.238509] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 999.422678] env[61006]: DEBUG nova.network.neutron [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.666402] env[61006]: DEBUG nova.scheduler.client.report [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.813349] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.813604] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.925649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.925991] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Instance network_info: |[{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 999.926453] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:08:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdb40136-9153-4611-9dc5-7bde778a2b5a', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.934288] env[61006]: DEBUG oslo.service.loopingcall [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.934513] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.934750] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25add825-c060-4b5c-930e-517f1ba19b10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.954539] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.954539] env[61006]: value = "task-1337473" [ 999.954539] env[61006]: _type = "Task" [ 999.954539] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.961940] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337473, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.039084] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.039399] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.039627] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.039822] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.040021] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.042732] env[61006]: INFO nova.compute.manager [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Terminating instance [ 1000.044685] env[61006]: DEBUG nova.compute.manager [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.044886] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.045796] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a92671c-a935-4c6e-8016-b9004d49162a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.053950] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.054214] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cce496d5-e8f2-4bde-9f9c-28b2199f87bd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.059986] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 1000.059986] env[61006]: value = "task-1337474" [ 1000.059986] env[61006]: _type = "Task" [ 1000.059986] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.069619] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337474, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.118034] env[61006]: DEBUG nova.compute.manager [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Received event network-changed-cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.118268] env[61006]: DEBUG nova.compute.manager [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Refreshing instance network info cache due to event network-changed-cdb40136-9153-4611-9dc5-7bde778a2b5a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.118527] env[61006]: DEBUG oslo_concurrency.lockutils [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.118710] env[61006]: DEBUG oslo_concurrency.lockutils [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.119015] env[61006]: DEBUG nova.network.neutron [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Refreshing network info cache for port cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.172073] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.174273] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.318s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.174485] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.193822] env[61006]: INFO nova.scheduler.client.report [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Deleted allocations for instance d027fd07-c409-46e5-857f-cdd2c0479f53 [ 1000.196359] env[61006]: INFO nova.scheduler.client.report [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted allocations for instance f048bd9c-048b-4439-bbbf-dfcee7f18d84 [ 1000.252443] env[61006]: DEBUG nova.objects.instance [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid 02408ca9-c580-444a-9608-a752146ca499 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.315596] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.464398] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337473, 'name': CreateVM_Task, 'duration_secs': 0.297464} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.464700] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.465256] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.465423] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.465743] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.465996] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9b626a3-cb6a-43f7-ac34-8dfa25c68870 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.470686] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1000.470686] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c00793-be10-620d-19dd-d8f2fedb541a" [ 1000.470686] env[61006]: _type = "Task" [ 1000.470686] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.477985] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c00793-be10-620d-19dd-d8f2fedb541a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.569967] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337474, 'name': PowerOffVM_Task, 'duration_secs': 0.165556} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.570339] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.570615] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.570884] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-add26ac2-1ae6-44b4-9bf2-b35ad2162caf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.632187] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.632418] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Deleting contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.632631] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleting the datastore file [datastore1] feb3c68c-4c03-40cd-89a5-3310bfcab359 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.632910] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac27d972-227c-482b-8316-341ee0f25169 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.641152] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for the task: (returnval){ [ 1000.641152] env[61006]: value = "task-1337476" [ 1000.641152] env[61006]: _type = "Task" [ 1000.641152] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.649587] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.706286] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f1fa3ed8-222c-44b1-8b9a-b4b5bedb5644 tempest-DeleteServersTestJSON-1116056315 tempest-DeleteServersTestJSON-1116056315-project-member] Lock "d027fd07-c409-46e5-857f-cdd2c0479f53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.695s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.707254] env[61006]: DEBUG oslo_concurrency.lockutils [None req-cf0f62a6-3086-4c53-ab48-a0cc9719a3a3 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "f048bd9c-048b-4439-bbbf-dfcee7f18d84" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.977s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.757821] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ecf2cb8e-da1f-4e67-8b49-394d9d2e8a21 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.273s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.840754] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.841019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.842581] env[61006]: INFO nova.compute.claims [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.981835] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c00793-be10-620d-19dd-d8f2fedb541a, 'name': SearchDatastore_Task, 'duration_secs': 0.011992} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.983320] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.983580] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.983825] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.983997] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.984203] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.987316] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d59c78f9-1793-4f40-a693-db8f3276109b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.999171] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.999360] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.002294] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de2b1a5a-2e0e-4b01-a9d5-00b16400b7b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.008118] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1001.008118] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52de6394-ce41-7ea5-9a55-970a83ce52c5" [ 1001.008118] env[61006]: _type = "Task" [ 1001.008118] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.017241] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52de6394-ce41-7ea5-9a55-970a83ce52c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.026300] env[61006]: DEBUG nova.network.neutron [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updated VIF entry in instance network info cache for port cdb40136-9153-4611-9dc5-7bde778a2b5a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.026662] env[61006]: DEBUG nova.network.neutron [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.153175] env[61006]: DEBUG oslo_vmware.api [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Task: {'id': task-1337476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.380011} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.153545] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.153751] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Deleted contents of the VM from datastore datastore1 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.153930] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.154127] env[61006]: INFO nova.compute.manager [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1001.154388] env[61006]: DEBUG oslo.service.loopingcall [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.154596] env[61006]: DEBUG nova.compute.manager [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.154690] env[61006]: DEBUG nova.network.neutron [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.492911] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.494335] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.494629] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.495068] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.495068] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.498925] env[61006]: INFO nova.compute.manager [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Terminating instance [ 1001.500958] env[61006]: DEBUG nova.compute.manager [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.501452] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.502458] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9efbeca-b737-43f5-8405-71a4ced5ea84 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.513868] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.514625] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85cc2c2d-8e80-421f-b852-c4288af3ad81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.519790] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52de6394-ce41-7ea5-9a55-970a83ce52c5, 'name': SearchDatastore_Task, 'duration_secs': 0.012143} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.521529] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 1001.521529] env[61006]: value = "task-1337477" [ 1001.521529] env[61006]: _type = "Task" [ 1001.521529] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.521736] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b92a1d6-2e4d-400b-ac12-45a988a56be1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.529308] env[61006]: DEBUG oslo_concurrency.lockutils [req-e8ba4dca-2317-4aa0-b524-abb19d18e30a req-07a9b329-bb24-4af0-b15d-d273dfce7339 service nova] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.533983] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.534281] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1001.534281] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b8824a-a294-996b-dceb-0f73efc6d971" [ 1001.534281] env[61006]: _type = "Task" [ 1001.534281] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.542569] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b8824a-a294-996b-dceb-0f73efc6d971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.984577] env[61006]: DEBUG nova.network.neutron [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.034710] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337477, 'name': PowerOffVM_Task, 'duration_secs': 0.278077} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.036040] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.036194] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.036925] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b532a90-3bed-49c9-8d21-a89bc6e78163 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.042287] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8de56d1-dcc5-4977-aeec-4e37f8cd32df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.051096] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e86fd3-e6fb-4705-8eba-b467634bd8ca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.054466] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52b8824a-a294-996b-dceb-0f73efc6d971, 'name': SearchDatastore_Task, 'duration_secs': 0.012043} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.054724] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.055040] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.055690] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db74f542-056a-48d7-bb8f-5171de4e6aa2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.089503] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86f703d-0275-4e34-8efa-abead55b004b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.094481] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1002.094481] env[61006]: value = "task-1337479" [ 1002.094481] env[61006]: _type = "Task" [ 1002.094481] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.100646] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e25353-99de-4672-91a1-4a88243d05ae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.108947] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.110963] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.111231] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.111438] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleting the datastore file [datastore2] bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.111732] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e011d72-0bc2-418a-80ea-ccc6442fd3d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.123057] env[61006]: DEBUG nova.compute.provider_tree [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.128571] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for the task: (returnval){ [ 1002.128571] env[61006]: value = "task-1337480" [ 1002.128571] env[61006]: _type = "Task" [ 1002.128571] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.138926] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.146972] env[61006]: DEBUG nova.compute.manager [req-97e05de8-5900-452b-a752-048f2e8272ba req-f9167ccf-6af0-448a-b971-85714d5add65 service nova] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Received event network-vif-deleted-368436a2-9db8-429d-9f98-d4f5baf42930 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.395243] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.396144] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.489681] env[61006]: INFO nova.compute.manager [-] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Took 1.33 seconds to deallocate network for instance. [ 1002.603332] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49582} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.603692] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.603875] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.604173] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdafd7e1-a276-4c42-9f70-42676cb88679 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.610609] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1002.610609] env[61006]: value = "task-1337481" [ 1002.610609] env[61006]: _type = "Task" [ 1002.610609] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.624562] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337481, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.625427] env[61006]: DEBUG nova.scheduler.client.report [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.637602] env[61006]: DEBUG oslo_vmware.api [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Task: {'id': task-1337480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.452471} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.637602] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.637679] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.639825] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.639825] env[61006]: INFO nova.compute.manager [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1002.639825] env[61006]: DEBUG oslo.service.loopingcall [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.639825] env[61006]: DEBUG nova.compute.manager [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.639825] env[61006]: DEBUG nova.network.neutron [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.858569] env[61006]: DEBUG oslo_concurrency.lockutils [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.859292] env[61006]: DEBUG oslo_concurrency.lockutils [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.898847] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1002.996859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.120015] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337481, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066148} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.120688] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.121064] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91af160a-b866-4720-baa6-e74d3bbbe31e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.134349] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.134847] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.146448] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.146994] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.150s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.147232] env[61006]: DEBUG nova.objects.instance [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lazy-loading 'resources' on Instance uuid feb3c68c-4c03-40cd-89a5-3310bfcab359 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.148563] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d43b0cf-c58b-469e-89a0-8135047a60e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.168638] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1003.168638] env[61006]: value = "task-1337482" [ 1003.168638] env[61006]: _type = "Task" [ 1003.168638] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.176892] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.343770] env[61006]: DEBUG nova.network.neutron [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.362389] env[61006]: INFO nova.compute.manager [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Detaching volume d7571e01-7903-40f1-8588-89faa7a25ab9 [ 1003.399165] env[61006]: INFO nova.virt.block_device [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Attempting to driver detach volume d7571e01-7903-40f1-8588-89faa7a25ab9 from mountpoint /dev/sdb [ 1003.399499] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1003.399713] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285404', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'name': 'volume-d7571e01-7903-40f1-8588-89faa7a25ab9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a546f17-2fb8-4b99-9001-98cc6fe76837', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'serial': 'd7571e01-7903-40f1-8588-89faa7a25ab9'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1003.400635] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6acbd45-855f-4aac-9cc9-10355e172e19 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.427927] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.428825] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9895a1f-198e-4bd5-ab43-5e7ccb268d0c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.441871] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376daf54-6f5c-49fa-b5c1-2f651e5c974f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.463351] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9c4b5a-2e3f-4282-81c3-38972e7a5668 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.478460] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] The volume has not been displaced from its original location: [datastore1] volume-d7571e01-7903-40f1-8588-89faa7a25ab9/volume-d7571e01-7903-40f1-8588-89faa7a25ab9.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1003.483748] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.484072] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7268b574-ccc5-44d6-bfee-2ecfbfc5b6b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.501559] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1003.501559] env[61006]: value = "task-1337483" [ 1003.501559] env[61006]: _type = "Task" [ 1003.501559] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.510103] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.648544] env[61006]: DEBUG nova.compute.utils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.650383] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1003.650647] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1003.680546] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337482, 'name': ReconfigVM_Task, 'duration_secs': 0.288131} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.680845] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.681492] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eefbd882-fad8-41db-afd4-8c09f0dfc306 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.688845] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1003.688845] env[61006]: value = "task-1337484" [ 1003.688845] env[61006]: _type = "Task" [ 1003.688845] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.696722] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337484, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.720097] env[61006]: DEBUG nova.policy [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1003.802372] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c07c034-7f85-4ef6-966c-d3bb4620395d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.809463] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2743cff2-40a6-453d-aefb-c4278cb2e2fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.840875] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f58c5b4-f05c-4cec-9e9d-8e410f9734d2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.847569] env[61006]: INFO nova.compute.manager [-] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Took 1.21 seconds to deallocate network for instance. [ 1003.850430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f56d3bc-d9a9-4f07-a268-8a194f502d28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.865721] env[61006]: DEBUG nova.compute.provider_tree [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.011467] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337483, 'name': ReconfigVM_Task, 'duration_secs': 0.228109} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.012121] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Successfully created port: 446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.013950] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1004.019739] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfb96732-3575-4cc2-87c8-ea8fdc2998dc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.035450] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1004.035450] env[61006]: value = "task-1337485" [ 1004.035450] env[61006]: _type = "Task" [ 1004.035450] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.043781] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337485, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.156444] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.171981] env[61006]: DEBUG nova.compute.manager [req-ad752e66-738b-49f7-af18-c2ab441a9356 req-f64f8020-e5f7-4ba1-9f65-f8d3aab10ca6 service nova] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Received event network-vif-deleted-58bb0db5-4669-4185-8d20-b5a77724df9b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.199778] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337484, 'name': Rename_Task, 'duration_secs': 0.136102} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.200090] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.200358] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ecd9d2f-56c0-451d-9da7-74aeec7efd6e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.207292] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1004.207292] env[61006]: value = "task-1337486" [ 1004.207292] env[61006]: _type = "Task" [ 1004.207292] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.215258] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337486, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.358281] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.368449] env[61006]: DEBUG nova.scheduler.client.report [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.547302] env[61006]: DEBUG oslo_vmware.api [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337485, 'name': ReconfigVM_Task, 'duration_secs': 0.122848} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.547607] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285404', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'name': 'volume-d7571e01-7903-40f1-8588-89faa7a25ab9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a546f17-2fb8-4b99-9001-98cc6fe76837', 'attached_at': '', 'detached_at': '', 'volume_id': 'd7571e01-7903-40f1-8588-89faa7a25ab9', 'serial': 'd7571e01-7903-40f1-8588-89faa7a25ab9'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1004.728538] env[61006]: DEBUG oslo_vmware.api [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337486, 'name': PowerOnVM_Task, 'duration_secs': 0.451785} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.728846] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.729063] env[61006]: INFO nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Took 6.82 seconds to spawn the instance on the hypervisor. [ 1004.729250] env[61006]: DEBUG nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.730054] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415d5096-e4c0-41c1-9bac-d64309287d06 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.873695] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.876015] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.448s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.877777] env[61006]: INFO nova.compute.claims [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.900835] env[61006]: INFO nova.scheduler.client.report [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Deleted allocations for instance feb3c68c-4c03-40cd-89a5-3310bfcab359 [ 1005.104376] env[61006]: DEBUG nova.objects.instance [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'flavor' on Instance uuid 1a546f17-2fb8-4b99-9001-98cc6fe76837 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.166597] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.206626] env[61006]: DEBUG nova.virt.hardware [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.207916] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de941314-a3be-40d5-8484-55c298250a47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.215728] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4f11db-78d9-4b04-80a0-2b3c898b98c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.249539] env[61006]: INFO nova.compute.manager [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Took 13.15 seconds to build instance. [ 1005.411706] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c7b85c84-40f9-42b5-b6e4-56bc865fac0a tempest-ServerDiskConfigTestJSON-1199123773 tempest-ServerDiskConfigTestJSON-1199123773-project-member] Lock "feb3c68c-4c03-40cd-89a5-3310bfcab359" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.371s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.751706] env[61006]: DEBUG oslo_concurrency.lockutils [None req-064698c3-7d3e-43b1-b77e-3669a63ef6c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.659s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.021125] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Successfully updated port: 446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.097412] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5aca55-fb8a-4366-a26c-288d553fd9ed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.104965] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c061bf0b-0b9c-43ef-b736-27b01824320d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.138859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-27c7d3f0-1878-4044-a078-d6ca9ebc340b tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.280s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.140977] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4aee6e8-8a0d-4504-a22b-240d0a7f693b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.149405] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17548b12-4c80-4397-a070-3fc1b2e7a683 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.166519] env[61006]: DEBUG nova.compute.provider_tree [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.210322] env[61006]: DEBUG nova.compute.manager [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Received event network-vif-plugged-446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.210322] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Acquiring lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.210322] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.210322] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.210322] env[61006]: DEBUG nova.compute.manager [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] No waiting events found dispatching network-vif-plugged-446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.211062] env[61006]: WARNING nova.compute.manager [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Received unexpected event network-vif-plugged-446e2c42-a832-49df-b34e-c50389d59cc4 for instance with vm_state building and task_state spawning. [ 1006.211354] env[61006]: DEBUG nova.compute.manager [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Received event network-changed-446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.211669] env[61006]: DEBUG nova.compute.manager [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Refreshing instance network info cache due to event network-changed-446e2c42-a832-49df-b34e-c50389d59cc4. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.211963] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Acquiring lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.212227] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Acquired lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.212478] env[61006]: DEBUG nova.network.neutron [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Refreshing network info cache for port 446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.359473] env[61006]: DEBUG nova.compute.manager [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Received event network-changed-cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.360979] env[61006]: DEBUG nova.compute.manager [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Refreshing instance network info cache due to event network-changed-cdb40136-9153-4611-9dc5-7bde778a2b5a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.360979] env[61006]: DEBUG oslo_concurrency.lockutils [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.360979] env[61006]: DEBUG oslo_concurrency.lockutils [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.360979] env[61006]: DEBUG nova.network.neutron [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Refreshing network info cache for port cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.527011] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.670326] env[61006]: DEBUG nova.scheduler.client.report [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.784192] env[61006]: DEBUG nova.network.neutron [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1006.904834] env[61006]: DEBUG nova.network.neutron [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.915091] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "9486ebfc-4717-4c97-8b44-ac922737bc78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.915091] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.117015] env[61006]: DEBUG nova.network.neutron [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updated VIF entry in instance network info cache for port cdb40136-9153-4611-9dc5-7bde778a2b5a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.117015] env[61006]: DEBUG nova.network.neutron [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.176382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.176951] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.184088] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.824s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.184088] env[61006]: DEBUG nova.objects.instance [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lazy-loading 'resources' on Instance uuid bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.265900] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.266246] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.266903] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.267189] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.267424] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.269742] env[61006]: INFO nova.compute.manager [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Terminating instance [ 1007.271823] env[61006]: DEBUG nova.compute.manager [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.272083] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.273091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5f3df9-cfcc-4b42-a669-be535abc464e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.281392] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.281699] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f94450a3-8b3b-4f7f-af05-fe280778ed2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.288393] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1007.288393] env[61006]: value = "task-1337487" [ 1007.288393] env[61006]: _type = "Task" [ 1007.288393] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.296753] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.415607] env[61006]: DEBUG oslo_concurrency.lockutils [req-2d9b014c-fc95-47f6-95bb-f9fa7e5f1589 req-0d81e61b-9a42-47b3-ba7e-d19bf7550f14 service nova] Releasing lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.416124] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.420022] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.420022] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.622850] env[61006]: DEBUG oslo_concurrency.lockutils [req-c82af0df-2d14-4ed8-b4a1-dfdcd7cbfa85 req-3816b9c6-3cff-43fd-995e-34cecb3d2386 service nova] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.690448] env[61006]: DEBUG nova.compute.utils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.692112] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.692185] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1007.739535] env[61006]: DEBUG nova.policy [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7c04e3804464cdc8c2925f7bbcf5369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4e87e7cb3f2444a8cc7bb7e65a99757', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1007.805983] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337487, 'name': PowerOffVM_Task, 'duration_secs': 0.186057} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.806227] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.806227] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.806456] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3df5c85-4e52-48cd-a938-6b582ba8eebe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.865485] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12401f60-b722-470a-b086-cb08ea6b8f05 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.873267] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d698cb-acf7-4577-9d38-1f6b20594e12 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.877686] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.877911] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.878110] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleting the datastore file [datastore2] 1a546f17-2fb8-4b99-9001-98cc6fe76837 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.878709] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-490240a7-529e-46f8-8546-0f13f2704890 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.909050] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e7b5b1-8483-46d4-97c2-fc93d1798c6c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.911720] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1007.911720] env[61006]: value = "task-1337489" [ 1007.911720] env[61006]: _type = "Task" [ 1007.911720] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.918855] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849e6128-85fe-4e79-9c49-936cfcd8d18d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.928990] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.942357] env[61006]: DEBUG nova.compute.provider_tree [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.943911] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.961877] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.070297] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Successfully created port: 4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.162414] env[61006]: DEBUG nova.network.neutron [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Updating instance_info_cache with network_info: [{"id": "446e2c42-a832-49df-b34e-c50389d59cc4", "address": "fa:16:3e:fc:79:6e", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446e2c42-a8", "ovs_interfaceid": "446e2c42-a832-49df-b34e-c50389d59cc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.195092] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.422139] env[61006]: DEBUG oslo_vmware.api [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360947} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.422448] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.422810] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.423412] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.423412] env[61006]: INFO nova.compute.manager [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1008.423533] env[61006]: DEBUG oslo.service.loopingcall [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.423713] env[61006]: DEBUG nova.compute.manager [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.423811] env[61006]: DEBUG nova.network.neutron [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.446688] env[61006]: DEBUG nova.scheduler.client.report [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.665440] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-a36c3e9b-83b9-434e-8fe1-601d389a8cf2" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.665752] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance network_info: |[{"id": "446e2c42-a832-49df-b34e-c50389d59cc4", "address": "fa:16:3e:fc:79:6e", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446e2c42-a8", "ovs_interfaceid": "446e2c42-a832-49df-b34e-c50389d59cc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.666223] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:79:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '446e2c42-a832-49df-b34e-c50389d59cc4', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.674375] env[61006]: DEBUG oslo.service.loopingcall [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.674607] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.674842] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbf8bbb8-bc4a-42e0-a202-a00422ecdde4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.695627] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.695627] env[61006]: value = "task-1337490" [ 1008.695627] env[61006]: _type = "Task" [ 1008.695627] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.702914] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337490, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.854872] env[61006]: DEBUG nova.compute.manager [req-3cf886b3-0693-4380-875a-dfc296de4a7b req-d509a7a6-6cff-49fa-a56d-92f63a466072 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Received event network-vif-deleted-43c495ff-c9d9-46fd-93e5-5c92d5f01d1f {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.855141] env[61006]: INFO nova.compute.manager [req-3cf886b3-0693-4380-875a-dfc296de4a7b req-d509a7a6-6cff-49fa-a56d-92f63a466072 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Neutron deleted interface 43c495ff-c9d9-46fd-93e5-5c92d5f01d1f; detaching it from the instance and deleting it from the info cache [ 1008.855395] env[61006]: DEBUG nova.network.neutron [req-3cf886b3-0693-4380-875a-dfc296de4a7b req-d509a7a6-6cff-49fa-a56d-92f63a466072 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.950859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.954346] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.010s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.958196] env[61006]: INFO nova.compute.claims [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.980790] env[61006]: INFO nova.scheduler.client.report [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Deleted allocations for instance bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a [ 1009.205946] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337490, 'name': CreateVM_Task, 'duration_secs': 0.304446} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.206144] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.206829] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.207009] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.207345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.207601] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2810e21c-570f-4355-a250-038a90be6ce6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.210704] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.215789] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1009.215789] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3d6b4-ee74-c3f9-069f-a29095770ddd" [ 1009.215789] env[61006]: _type = "Task" [ 1009.215789] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.225059] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3d6b4-ee74-c3f9-069f-a29095770ddd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.236408] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.236645] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.236805] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.236987] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.237152] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.237303] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.237508] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.237672] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.237843] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.238012] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.238202] env[61006]: DEBUG nova.virt.hardware [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.238960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0c8ec5-862a-4f45-8718-ed9fe6751a35 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.246322] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a94134-1044-482e-a9cb-020cd189ae4f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.331994] env[61006]: DEBUG nova.network.neutron [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.357860] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2a87e27-6ec7-4fd8-8277-c2081e3b3c0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.366955] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb868cd5-3412-483a-aeda-5b2ad8ce2028 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.394740] env[61006]: DEBUG nova.compute.manager [req-3cf886b3-0693-4380-875a-dfc296de4a7b req-d509a7a6-6cff-49fa-a56d-92f63a466072 service nova] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Detach interface failed, port_id=43c495ff-c9d9-46fd-93e5-5c92d5f01d1f, reason: Instance 1a546f17-2fb8-4b99-9001-98cc6fe76837 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1009.490924] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f42f2ee0-a52e-4e1a-b4ef-af83b8a3fd77 tempest-AttachInterfacesTestJSON-1276667898 tempest-AttachInterfacesTestJSON-1276667898-project-member] Lock "bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.997s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.506577] env[61006]: DEBUG nova.compute.manager [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Received event network-vif-plugged-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.506771] env[61006]: DEBUG oslo_concurrency.lockutils [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] Acquiring lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.507045] env[61006]: DEBUG oslo_concurrency.lockutils [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.507232] env[61006]: DEBUG oslo_concurrency.lockutils [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.507404] env[61006]: DEBUG nova.compute.manager [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] No waiting events found dispatching network-vif-plugged-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1009.507571] env[61006]: WARNING nova.compute.manager [req-e9dd0745-c073-4558-b3e5-e688394542fc req-b020563c-0259-4871-be40-a64a19cf5e22 service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Received unexpected event network-vif-plugged-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 for instance with vm_state building and task_state spawning. [ 1009.596882] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Successfully updated port: 4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.730627] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c3d6b4-ee74-c3f9-069f-a29095770ddd, 'name': SearchDatastore_Task, 'duration_secs': 0.00894} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.731014] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.731325] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.731639] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.731845] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.732100] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.732434] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3a5fe6c-1c4c-417f-864c-d4ec04681603 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.741777] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.742107] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.742986] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-655d247a-4aac-4367-95f3-043c1211c662 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.749738] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1009.749738] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52783515-8b58-acbf-d6a9-8c8a82cd8251" [ 1009.749738] env[61006]: _type = "Task" [ 1009.749738] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.759520] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52783515-8b58-acbf-d6a9-8c8a82cd8251, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.833930] env[61006]: INFO nova.compute.manager [-] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Took 1.41 seconds to deallocate network for instance. [ 1010.085410] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c652b7f-fc1b-4ce7-a1c5-8839015f672c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.093936] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13fb75a4-4285-4892-a79a-5887875d951e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.099477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.099616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.099764] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.124771] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d202c22c-c733-470b-8487-421ccf69fb47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.132224] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1f83ee-390a-47c5-9423-ea2a5fe640d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.148261] env[61006]: DEBUG nova.compute.provider_tree [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.260015] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52783515-8b58-acbf-d6a9-8c8a82cd8251, 'name': SearchDatastore_Task, 'duration_secs': 0.008244} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.260795] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c570c501-6dd8-4878-b872-3d0c6b35064f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.265789] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1010.265789] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c507c5-440e-8e0d-d2a1-acb4b631bfa4" [ 1010.265789] env[61006]: _type = "Task" [ 1010.265789] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.275189] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c507c5-440e-8e0d-d2a1-acb4b631bfa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.341345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.651131] env[61006]: DEBUG nova.scheduler.client.report [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.655451] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.776271] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c507c5-440e-8e0d-d2a1-acb4b631bfa4, 'name': SearchDatastore_Task, 'duration_secs': 0.008904} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.776536] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.776802] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a36c3e9b-83b9-434e-8fe1-601d389a8cf2/a36c3e9b-83b9-434e-8fe1-601d389a8cf2.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.777085] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba779c33-7e88-46b7-9538-2204ffa60004 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.783600] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1010.783600] env[61006]: value = "task-1337491" [ 1010.783600] env[61006]: _type = "Task" [ 1010.783600] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.791391] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.885846] env[61006]: DEBUG nova.network.neutron [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updating instance_info_cache with network_info: [{"id": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "address": "fa:16:3e:b9:d9:aa", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f0a0867-f2", "ovs_interfaceid": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.156460] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.157096] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.160871] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.820s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.161557] env[61006]: DEBUG nova.objects.instance [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'resources' on Instance uuid 1a546f17-2fb8-4b99-9001-98cc6fe76837 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.294260] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475219} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.294577] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a36c3e9b-83b9-434e-8fe1-601d389a8cf2/a36c3e9b-83b9-434e-8fe1-601d389a8cf2.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.294795] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.295092] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad885a99-e557-4d38-9533-c181a7c53ec2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.301896] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1011.301896] env[61006]: value = "task-1337492" [ 1011.301896] env[61006]: _type = "Task" [ 1011.301896] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.310909] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337492, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.388925] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.389329] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Instance network_info: |[{"id": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "address": "fa:16:3e:b9:d9:aa", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f0a0867-f2", "ovs_interfaceid": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1011.389801] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:d9:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4f0a0867-f2dd-4b43-a42c-05bce1275bb6', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.397657] env[61006]: DEBUG oslo.service.loopingcall [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.397898] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.398147] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8fffed4-82f4-44a0-8eb3-9cd380e76ca6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.418562] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.418562] env[61006]: value = "task-1337493" [ 1011.418562] env[61006]: _type = "Task" [ 1011.418562] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.428365] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337493, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.543653] env[61006]: DEBUG nova.compute.manager [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Received event network-changed-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.543791] env[61006]: DEBUG nova.compute.manager [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Refreshing instance network info cache due to event network-changed-4f0a0867-f2dd-4b43-a42c-05bce1275bb6. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.544020] env[61006]: DEBUG oslo_concurrency.lockutils [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] Acquiring lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.544360] env[61006]: DEBUG oslo_concurrency.lockutils [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] Acquired lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.544508] env[61006]: DEBUG nova.network.neutron [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Refreshing network info cache for port 4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.662562] env[61006]: DEBUG nova.compute.utils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.665076] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.665377] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.736678] env[61006]: DEBUG nova.policy [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3bd50700ed14e45bff0fda08e5c7e8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bf43cb0133740e9ad70d2c2af3ddd56', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.814536] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337492, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060112} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.815584] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.816352] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1852e360-dce9-40b5-bade-b684bddbcca5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.819219] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e81eb8-f492-4167-8885-1ebf8c1da9dc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.834428] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb16e74b-dcbc-438c-a986-67146d7b39ee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.846979] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] a36c3e9b-83b9-434e-8fe1-601d389a8cf2/a36c3e9b-83b9-434e-8fe1-601d389a8cf2.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.847285] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01845f6e-2047-48f0-9b07-39c63dcab783 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.894044] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f0e9f8-5d36-4ce4-bff4-3e1f2cbe1242 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.896876] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1011.896876] env[61006]: value = "task-1337494" [ 1011.896876] env[61006]: _type = "Task" [ 1011.896876] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.904517] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffcd2a5-969b-4395-a71f-1aaca6acb24e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.911143] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337494, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.920418] env[61006]: DEBUG nova.compute.provider_tree [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.929556] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337493, 'name': CreateVM_Task, 'duration_secs': 0.282737} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.929719] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.930394] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.930563] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.930875] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.931137] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac215fa5-eece-408d-9c24-ef11f394a168 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.935760] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1011.935760] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52785019-6f95-7baa-f6a1-a2447aa568e7" [ 1011.935760] env[61006]: _type = "Task" [ 1011.935760] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.944860] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52785019-6f95-7baa-f6a1-a2447aa568e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.049545] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Successfully created port: 3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.165830] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.372927] env[61006]: DEBUG nova.network.neutron [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updated VIF entry in instance network info cache for port 4f0a0867-f2dd-4b43-a42c-05bce1275bb6. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.373439] env[61006]: DEBUG nova.network.neutron [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updating instance_info_cache with network_info: [{"id": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "address": "fa:16:3e:b9:d9:aa", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f0a0867-f2", "ovs_interfaceid": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.407850] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337494, 'name': ReconfigVM_Task, 'duration_secs': 0.263059} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.408148] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Reconfigured VM instance instance-00000066 to attach disk [datastore2] a36c3e9b-83b9-434e-8fe1-601d389a8cf2/a36c3e9b-83b9-434e-8fe1-601d389a8cf2.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.408825] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-407a4aa2-ef71-4c43-8d92-4b800869f8c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.415528] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1012.415528] env[61006]: value = "task-1337495" [ 1012.415528] env[61006]: _type = "Task" [ 1012.415528] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.424260] env[61006]: DEBUG nova.scheduler.client.report [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.427934] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337495, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.450381] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52785019-6f95-7baa-f6a1-a2447aa568e7, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.450773] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.451027] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.451275] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.451421] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.451600] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.451879] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd218b9f-2f99-4d5f-a8e6-79da3a05b69f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.460506] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.460726] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.461648] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7554f8fa-90c7-4318-a2f7-a6a61b56b7c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.466830] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1012.466830] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae42fa-a840-e3d6-be26-3f11a00af4f1" [ 1012.466830] env[61006]: _type = "Task" [ 1012.466830] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.475019] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae42fa-a840-e3d6-be26-3f11a00af4f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.877417] env[61006]: DEBUG oslo_concurrency.lockutils [req-821d5a59-73f9-4edd-a6e1-f6816f34c68e req-29eeae93-27fa-4b31-9878-519e9085029e service nova] Releasing lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.925286] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337495, 'name': Rename_Task, 'duration_secs': 0.140891} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.925578] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.925828] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9173c514-7b55-4089-b9fd-bb22acf34cec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.929141] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.932352] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1012.932352] env[61006]: value = "task-1337496" [ 1012.932352] env[61006]: _type = "Task" [ 1012.932352] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.939693] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337496, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.957770] env[61006]: INFO nova.scheduler.client.report [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted allocations for instance 1a546f17-2fb8-4b99-9001-98cc6fe76837 [ 1012.980222] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae42fa-a840-e3d6-be26-3f11a00af4f1, 'name': SearchDatastore_Task, 'duration_secs': 0.008018} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.980222] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e31f0337-c68a-44c0-b588-ada18fccdb73 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.984200] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1012.984200] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae3c69-c93b-8fad-4a0e-3eb3e3d31d03" [ 1012.984200] env[61006]: _type = "Task" [ 1012.984200] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.993726] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae3c69-c93b-8fad-4a0e-3eb3e3d31d03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.176068] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.203255] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.203576] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.203816] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.204043] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.204202] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.204354] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.204564] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.204727] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.204897] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.205083] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.205269] env[61006]: DEBUG nova.virt.hardware [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.206170] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca95a28-401f-4cb9-9f82-d335e848f792 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.214979] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292954b7-0caa-4a61-a9d4-9a0bf8ad819a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.442500] env[61006]: DEBUG oslo_vmware.api [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337496, 'name': PowerOnVM_Task, 'duration_secs': 0.482413} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.442923] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.443151] env[61006]: INFO nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Took 8.28 seconds to spawn the instance on the hypervisor. [ 1013.443336] env[61006]: DEBUG nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1013.444097] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b3fab1-a6f4-4343-8009-4510482aa553 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.465439] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c310dd8f-f2dd-4936-9cde-278010682fe0 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "1a546f17-2fb8-4b99-9001-98cc6fe76837" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.199s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.496018] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ae3c69-c93b-8fad-4a0e-3eb3e3d31d03, 'name': SearchDatastore_Task, 'duration_secs': 0.009057} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.496299] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.496558] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 5c1c8651-cb82-49c7-9e0d-fe3960fc1633/5c1c8651-cb82-49c7-9e0d-fe3960fc1633.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.496857] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6438bfe6-4fc1-4f99-8b1d-f886b1898f98 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.504433] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1013.504433] env[61006]: value = "task-1337497" [ 1013.504433] env[61006]: _type = "Task" [ 1013.504433] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.514022] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.679462] env[61006]: DEBUG nova.compute.manager [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Received event network-vif-plugged-3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.679690] env[61006]: DEBUG oslo_concurrency.lockutils [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] Acquiring lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.679894] env[61006]: DEBUG oslo_concurrency.lockutils [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.680074] env[61006]: DEBUG oslo_concurrency.lockutils [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.680250] env[61006]: DEBUG nova.compute.manager [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] No waiting events found dispatching network-vif-plugged-3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1013.680418] env[61006]: WARNING nova.compute.manager [req-166eed8f-9ef3-4216-808d-d1d8972b0117 req-96d081ca-3dbc-4d0a-8934-b835d01c5f40 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Received unexpected event network-vif-plugged-3ddc0fab-3451-455f-9b83-3b55726b6d6b for instance with vm_state building and task_state spawning. [ 1013.919812] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Successfully updated port: 3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.960287] env[61006]: INFO nova.compute.manager [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Took 13.14 seconds to build instance. [ 1014.015578] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495168} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.016250] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 5c1c8651-cb82-49c7-9e0d-fe3960fc1633/5c1c8651-cb82-49c7-9e0d-fe3960fc1633.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.016250] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.016407] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7682e7c5-61f1-4d77-99cd-88e6c73b5074 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.022930] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1014.022930] env[61006]: value = "task-1337499" [ 1014.022930] env[61006]: _type = "Task" [ 1014.022930] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.030943] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337499, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.425219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.425515] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquired lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.425563] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.462856] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2113ef1d-7c92-4213-af2c-de9c1e3b1abf tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.649s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.533062] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337499, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064012} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.533348] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.534169] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3a4956-f185-4311-98fb-d211824ec1b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.557137] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 5c1c8651-cb82-49c7-9e0d-fe3960fc1633/5c1c8651-cb82-49c7-9e0d-fe3960fc1633.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.557416] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c577a01-d407-4407-b231-5e84c0b1712d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.576640] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1014.576640] env[61006]: value = "task-1337500" [ 1014.576640] env[61006]: _type = "Task" [ 1014.576640] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.584073] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337500, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.820704] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.820978] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.821219] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.821413] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.821586] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.824019] env[61006]: INFO nova.compute.manager [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Terminating instance [ 1014.825838] env[61006]: DEBUG nova.compute.manager [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.826051] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.826923] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd0442e-691e-4a36-9e34-83e062b35743 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.834297] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.834516] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cbdeb5d-d5e2-4c48-b30d-a85d3be2a677 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.840213] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1014.840213] env[61006]: value = "task-1337501" [ 1014.840213] env[61006]: _type = "Task" [ 1014.840213] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.847084] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.956942] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1015.086748] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337500, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.101529] env[61006]: DEBUG nova.network.neutron [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Updating instance_info_cache with network_info: [{"id": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "address": "fa:16:3e:39:71:10", "network": {"id": "3fd4d914-ad2d-4741-8eb8-02b97a509da7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2000423955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bf43cb0133740e9ad70d2c2af3ddd56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddc0fab-34", "ovs_interfaceid": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.350751] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337501, 'name': PowerOffVM_Task, 'duration_secs': 0.177844} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.351026] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.351205] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.351459] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ab6602e-110d-48df-aec9-55a48e2a0bae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.416133] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.416361] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.416547] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore2] a36c3e9b-83b9-434e-8fe1-601d389a8cf2 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.416810] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-444fb2ef-0104-47dc-911c-2e23ca76240c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.423846] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1015.423846] env[61006]: value = "task-1337503" [ 1015.423846] env[61006]: _type = "Task" [ 1015.423846] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.432239] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.587704] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337500, 'name': ReconfigVM_Task, 'duration_secs': 0.538934} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.588010] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 5c1c8651-cb82-49c7-9e0d-fe3960fc1633/5c1c8651-cb82-49c7-9e0d-fe3960fc1633.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.588705] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-275327df-88d7-4c74-873f-d58c9fb09913 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.595043] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1015.595043] env[61006]: value = "task-1337504" [ 1015.595043] env[61006]: _type = "Task" [ 1015.595043] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.603359] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337504, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.604859] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Releasing lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.605164] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Instance network_info: |[{"id": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "address": "fa:16:3e:39:71:10", "network": {"id": "3fd4d914-ad2d-4741-8eb8-02b97a509da7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2000423955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bf43cb0133740e9ad70d2c2af3ddd56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddc0fab-34", "ovs_interfaceid": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.605525] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:71:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ddc0fab-3451-455f-9b83-3b55726b6d6b', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.612711] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Creating folder: Project (8bf43cb0133740e9ad70d2c2af3ddd56). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.612959] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a809715-14bb-486f-bc27-e0caefce4e2d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.621402] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Created folder: Project (8bf43cb0133740e9ad70d2c2af3ddd56) in parent group-v285275. [ 1015.621580] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Creating folder: Instances. Parent ref: group-v285421. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.621787] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3d6dd97-9f27-4b8d-b641-685159728d42 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.630984] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Created folder: Instances in parent group-v285421. [ 1015.631224] env[61006]: DEBUG oslo.service.loopingcall [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.631406] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.631595] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-499a121a-0347-4200-926d-d968e87777f3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.650008] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.650008] env[61006]: value = "task-1337507" [ 1015.650008] env[61006]: _type = "Task" [ 1015.650008] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.659841] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337507, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.709492] env[61006]: DEBUG nova.compute.manager [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Received event network-changed-3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.709825] env[61006]: DEBUG nova.compute.manager [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Refreshing instance network info cache due to event network-changed-3ddc0fab-3451-455f-9b83-3b55726b6d6b. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.710237] env[61006]: DEBUG oslo_concurrency.lockutils [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] Acquiring lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.710537] env[61006]: DEBUG oslo_concurrency.lockutils [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] Acquired lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.710880] env[61006]: DEBUG nova.network.neutron [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Refreshing network info cache for port 3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.936116] env[61006]: DEBUG oslo_vmware.api [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177206} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.936293] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.936484] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.936657] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.936832] env[61006]: INFO nova.compute.manager [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1015.937090] env[61006]: DEBUG oslo.service.loopingcall [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.937332] env[61006]: DEBUG nova.compute.manager [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1015.937426] env[61006]: DEBUG nova.network.neutron [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.106290] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337504, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.161468] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337507, 'name': CreateVM_Task, 'duration_secs': 0.356389} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.161632] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1016.162320] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.162491] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.162855] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1016.163135] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-965cd896-42bd-4be0-8853-0d25bdc5c915 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.167757] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1016.167757] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5207d6fb-f238-397e-c0a7-1ad6ea4850d4" [ 1016.167757] env[61006]: _type = "Task" [ 1016.167757] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.176008] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5207d6fb-f238-397e-c0a7-1ad6ea4850d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.442418] env[61006]: DEBUG nova.network.neutron [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Updated VIF entry in instance network info cache for port 3ddc0fab-3451-455f-9b83-3b55726b6d6b. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.442879] env[61006]: DEBUG nova.network.neutron [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Updating instance_info_cache with network_info: [{"id": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "address": "fa:16:3e:39:71:10", "network": {"id": "3fd4d914-ad2d-4741-8eb8-02b97a509da7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2000423955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bf43cb0133740e9ad70d2c2af3ddd56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ddc0fab-34", "ovs_interfaceid": "3ddc0fab-3451-455f-9b83-3b55726b6d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.607309] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337504, 'name': Rename_Task, 'duration_secs': 0.600841} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.607656] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.607951] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11874027-5425-4750-b14d-ed97cd8d80f5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.614427] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1016.614427] env[61006]: value = "task-1337508" [ 1016.614427] env[61006]: _type = "Task" [ 1016.614427] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.621974] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337508, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.641401] env[61006]: DEBUG nova.network.neutron [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.677509] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5207d6fb-f238-397e-c0a7-1ad6ea4850d4, 'name': SearchDatastore_Task, 'duration_secs': 0.021137} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.677818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.678074] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.678318] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.678469] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.679055] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.679055] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4e965e6-2805-49e3-8f91-c113e495edac {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.687480] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.687662] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.688808] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf713e32-8449-4af6-8c49-8a155432c64c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.694066] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1016.694066] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a4e0b6-817a-7f20-8cee-cf774d320faf" [ 1016.694066] env[61006]: _type = "Task" [ 1016.694066] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.701494] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a4e0b6-817a-7f20-8cee-cf774d320faf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.945728] env[61006]: DEBUG oslo_concurrency.lockutils [req-810d0b27-a936-45c6-b759-d79670acd1a8 req-da05de9c-5d73-4b98-9f2e-a329c28c358a service nova] Releasing lock "refresh_cache-9486ebfc-4717-4c97-8b44-ac922737bc78" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.124897] env[61006]: DEBUG oslo_vmware.api [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337508, 'name': PowerOnVM_Task, 'duration_secs': 0.452348} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.125204] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.125408] env[61006]: INFO nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1017.125589] env[61006]: DEBUG nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.126397] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d716ed4c-16d3-4cbd-a4a2-d20e88deab13 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.143666] env[61006]: INFO nova.compute.manager [-] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Took 1.21 seconds to deallocate network for instance. [ 1017.203984] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a4e0b6-817a-7f20-8cee-cf774d320faf, 'name': SearchDatastore_Task, 'duration_secs': 0.008828} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.204737] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0440fe89-5201-470f-834d-f15e5318a38d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.209662] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1017.209662] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5225417e-30c3-6b1d-a93d-ba1e9d6c7779" [ 1017.209662] env[61006]: _type = "Task" [ 1017.209662] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.218854] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5225417e-30c3-6b1d-a93d-ba1e9d6c7779, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.642900] env[61006]: INFO nova.compute.manager [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Took 14.24 seconds to build instance. [ 1017.652932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.652932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.653167] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.653269] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.653926] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.655708] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.655937] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.656691] env[61006]: DEBUG nova.objects.instance [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid a36c3e9b-83b9-434e-8fe1-601d389a8cf2 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.658905] env[61006]: INFO nova.compute.manager [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Terminating instance [ 1017.660684] env[61006]: DEBUG nova.compute.manager [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.660884] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1017.661968] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f7a810-deaf-4b8e-8e6e-b944537d9f7d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.671342] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.671575] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40e27033-dbec-418c-9b80-2a3d74e05f81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.677868] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1017.677868] env[61006]: value = "task-1337509" [ 1017.677868] env[61006]: _type = "Task" [ 1017.677868] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.685548] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337509, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.719857] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5225417e-30c3-6b1d-a93d-ba1e9d6c7779, 'name': SearchDatastore_Task, 'duration_secs': 0.009407} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.720165] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.720424] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 9486ebfc-4717-4c97-8b44-ac922737bc78/9486ebfc-4717-4c97-8b44-ac922737bc78.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.720732] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-748d7875-dd22-40fd-b198-59c2981bf8a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.727417] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1017.727417] env[61006]: value = "task-1337510" [ 1017.727417] env[61006]: _type = "Task" [ 1017.727417] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.735616] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.737806] env[61006]: DEBUG nova.compute.manager [req-173b20ed-fbd6-4706-b24f-3c5c5bffe87e req-9430c1f4-7350-4526-bcb3-ea7bc0fa0fe8 service nova] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Received event network-vif-deleted-446e2c42-a832-49df-b34e-c50389d59cc4 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.145489] env[61006]: DEBUG oslo_concurrency.lockutils [None req-72f67c83-590d-4a7f-92a0-51c4e77c7e5a tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.750s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.191659] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337509, 'name': PowerOffVM_Task, 'duration_secs': 0.200909} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.191974] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.192166] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.192441] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-616e8a31-f667-4905-9257-56a33ebb65d9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.238483] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468114} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.238756] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 9486ebfc-4717-4c97-8b44-ac922737bc78/9486ebfc-4717-4c97-8b44-ac922737bc78.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.238973] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.239249] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-959057ed-1fb8-4809-b248-2470d56f9724 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.245751] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1018.245751] env[61006]: value = "task-1337512" [ 1018.245751] env[61006]: _type = "Task" [ 1018.245751] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.254388] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.260151] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.260354] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.260537] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleting the datastore file [datastore2] aedaa5d6-e0f2-492c-a14b-3254863e1f06 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.260797] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b01e28c-6353-41de-b33b-9ded539a9d28 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.266517] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1018.266517] env[61006]: value = "task-1337513" [ 1018.266517] env[61006]: _type = "Task" [ 1018.266517] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.276360] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.290776] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9be146-aa47-4452-a8aa-7d2c865c483c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.297292] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1160006a-f136-4299-8385-2a43be740d0c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.327882] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f06356-7334-4112-9a4b-6e670a9bca00 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.335065] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbb51da-4237-4cd5-b8bd-0febf5651072 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.348130] env[61006]: DEBUG nova.compute.provider_tree [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.758807] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071715} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.759176] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.760248] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc8bb8f-e0dc-4c1e-87b6-2856e533de1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.793544] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 9486ebfc-4717-4c97-8b44-ac922737bc78/9486ebfc-4717-4c97-8b44-ac922737bc78.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.796755] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2116af2-f7b2-43e7-8ef5-b2fa5f5a4697 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.815594] env[61006]: DEBUG oslo_vmware.api [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220619} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.816717] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.816918] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1018.817123] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1018.817308] env[61006]: INFO nova.compute.manager [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1018.817552] env[61006]: DEBUG oslo.service.loopingcall [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.817802] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1018.817802] env[61006]: value = "task-1337514" [ 1018.817802] env[61006]: _type = "Task" [ 1018.817802] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.817997] env[61006]: DEBUG nova.compute.manager [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1018.818115] env[61006]: DEBUG nova.network.neutron [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1018.827139] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337514, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.851444] env[61006]: DEBUG nova.scheduler.client.report [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.330482] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337514, 'name': ReconfigVM_Task, 'duration_secs': 0.249744} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.330798] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 9486ebfc-4717-4c97-8b44-ac922737bc78/9486ebfc-4717-4c97-8b44-ac922737bc78.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.331423] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbd4cb06-1aeb-4a69-9b4e-934d180f3c63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.337448] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1019.337448] env[61006]: value = "task-1337515" [ 1019.337448] env[61006]: _type = "Task" [ 1019.337448] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.344953] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337515, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.356844] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.381065] env[61006]: INFO nova.scheduler.client.report [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance a36c3e9b-83b9-434e-8fe1-601d389a8cf2 [ 1019.550537] env[61006]: DEBUG nova.network.neutron [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.775385] env[61006]: DEBUG nova.compute.manager [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Received event network-changed-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.775668] env[61006]: DEBUG nova.compute.manager [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Refreshing instance network info cache due to event network-changed-4f0a0867-f2dd-4b43-a42c-05bce1275bb6. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.775929] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] Acquiring lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.776161] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] Acquired lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.776320] env[61006]: DEBUG nova.network.neutron [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Refreshing network info cache for port 4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.847702] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337515, 'name': Rename_Task, 'duration_secs': 0.128925} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.848016] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.848296] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-257fea01-afcc-49c1-9f8f-bdd564fd39c4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.855651] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1019.855651] env[61006]: value = "task-1337516" [ 1019.855651] env[61006]: _type = "Task" [ 1019.855651] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.863745] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337516, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.889201] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ca6500d1-03e0-47a3-be1a-12f6581d1f05 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "a36c3e9b-83b9-434e-8fe1-601d389a8cf2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.068s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.053730] env[61006]: INFO nova.compute.manager [-] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Took 1.24 seconds to deallocate network for instance. [ 1020.364889] env[61006]: DEBUG oslo_vmware.api [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337516, 'name': PowerOnVM_Task, 'duration_secs': 0.447166} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.365199] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.365405] env[61006]: INFO nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Took 7.19 seconds to spawn the instance on the hypervisor. [ 1020.365587] env[61006]: DEBUG nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.366347] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2efdbb-6883-4ce9-bd34-466d3ad8e828 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.455345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.455601] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.493498] env[61006]: DEBUG nova.network.neutron [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updated VIF entry in instance network info cache for port 4f0a0867-f2dd-4b43-a42c-05bce1275bb6. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.493940] env[61006]: DEBUG nova.network.neutron [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updating instance_info_cache with network_info: [{"id": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "address": "fa:16:3e:b9:d9:aa", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4f0a0867-f2", "ovs_interfaceid": "4f0a0867-f2dd-4b43-a42c-05bce1275bb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.560711] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.561232] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.561657] env[61006]: DEBUG nova.objects.instance [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'resources' on Instance uuid aedaa5d6-e0f2-492c-a14b-3254863e1f06 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.742223] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.742529] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.882414] env[61006]: INFO nova.compute.manager [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Took 12.95 seconds to build instance. [ 1020.959077] env[61006]: INFO nova.compute.manager [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Detaching volume ce8406c1-edd0-4c87-b101-c9fc03c9897a [ 1020.994307] env[61006]: INFO nova.virt.block_device [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Attempting to driver detach volume ce8406c1-edd0-4c87-b101-c9fc03c9897a from mountpoint /dev/sdb [ 1020.994564] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1020.994756] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285400', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'name': 'volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '66c93148-b91a-4d22-84af-f410c8e10875', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'serial': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1020.995784] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5520c8d-f46d-4489-aa5e-6f5bc02baf80 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.998738] env[61006]: DEBUG oslo_concurrency.lockutils [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] Releasing lock "refresh_cache-5c1c8651-cb82-49c7-9e0d-fe3960fc1633" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.999566] env[61006]: DEBUG nova.compute.manager [req-1e63af74-c898-4f20-a1d0-90eca945340e req-52050b0d-bd1f-41f8-9c08-027d8e52eefa service nova] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Received event network-vif-deleted-63ee7dc4-4d60-4d0f-a9fc-6edc538fbeeb {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.018844] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6050e379-b220-4df4-bac1-5e7fb4d22c63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.026314] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c83dcf9-ac92-4456-a0f3-43ed22d4e75d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.046490] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94da974f-a68e-4803-add5-ef2efcb75cff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.061666] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] The volume has not been displaced from its original location: [datastore1] volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a/volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1021.066965] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1021.069553] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a43f9f6-8893-47f2-84bb-0d9eaf8d3801 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.088994] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1021.088994] env[61006]: value = "task-1337517" [ 1021.088994] env[61006]: _type = "Task" [ 1021.088994] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.096730] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337517, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.202837] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a8ae09-6d6f-43ff-abcb-e612c1c168f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.213207] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee397a08-198c-4273-b099-dcd980f08070 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.242613] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dbc87e-ccdd-4f78-8c59-f5841d411511 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.245397] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1021.253466] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5309f16c-825f-4cda-acac-61affa5e6330 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.266955] env[61006]: DEBUG nova.compute.provider_tree [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.283632] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "9486ebfc-4717-4c97-8b44-ac922737bc78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.383944] env[61006]: DEBUG oslo_concurrency.lockutils [None req-37604117-3036-40f8-9a66-62ad85d3a5b2 tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.469s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.384192] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.101s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.384422] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.384633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.384820] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.387455] env[61006]: INFO nova.compute.manager [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Terminating instance [ 1021.389246] env[61006]: DEBUG nova.compute.manager [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.389454] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.390278] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e85abaf-ec0a-48db-b711-6de49a41c25d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.397615] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.397843] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9691b80-cb31-4149-a3b4-f160b6c14e0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.403933] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1021.403933] env[61006]: value = "task-1337518" [ 1021.403933] env[61006]: _type = "Task" [ 1021.403933] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.411170] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337518, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.599545] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337517, 'name': ReconfigVM_Task, 'duration_secs': 0.467255} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.599846] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1021.604797] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0697e03c-0945-434b-9f6a-c862cb7dad30 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.620313] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1021.620313] env[61006]: value = "task-1337519" [ 1021.620313] env[61006]: _type = "Task" [ 1021.620313] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.628682] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337519, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.763617] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.769676] env[61006]: DEBUG nova.scheduler.client.report [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.915362] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337518, 'name': PowerOffVM_Task, 'duration_secs': 0.197522} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.915682] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.916213] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.916213] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8bbb11fc-d1bb-4ed2-8712-73d4b39ebf09 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.977050] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.977314] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.977507] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Deleting the datastore file [datastore2] 9486ebfc-4717-4c97-8b44-ac922737bc78 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.977778] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53234b92-cdd7-4ef1-bb88-8016dcf87c0a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.984239] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for the task: (returnval){ [ 1021.984239] env[61006]: value = "task-1337521" [ 1021.984239] env[61006]: _type = "Task" [ 1021.984239] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.991760] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337521, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.130049] env[61006]: DEBUG oslo_vmware.api [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337519, 'name': ReconfigVM_Task, 'duration_secs': 0.154934} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.130433] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285400', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'name': 'volume-ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '66c93148-b91a-4d22-84af-f410c8e10875', 'attached_at': '', 'detached_at': '', 'volume_id': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a', 'serial': 'ce8406c1-edd0-4c87-b101-c9fc03c9897a'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1022.274772] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.277059] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.514s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.278513] env[61006]: INFO nova.compute.claims [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1022.294152] env[61006]: INFO nova.scheduler.client.report [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted allocations for instance aedaa5d6-e0f2-492c-a14b-3254863e1f06 [ 1022.493995] env[61006]: DEBUG oslo_vmware.api [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Task: {'id': task-1337521, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172361} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.494301] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.494498] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.494685] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.494873] env[61006]: INFO nova.compute.manager [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1022.495133] env[61006]: DEBUG oslo.service.loopingcall [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.495336] env[61006]: DEBUG nova.compute.manager [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.495431] env[61006]: DEBUG nova.network.neutron [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.677805] env[61006]: DEBUG nova.objects.instance [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'flavor' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.777842] env[61006]: DEBUG nova.compute.manager [req-d06a084a-7029-4ba7-8629-04ca9b8d7cac req-5291f294-25b5-4f19-be66-0855d9f95887 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Received event network-vif-deleted-3ddc0fab-3451-455f-9b83-3b55726b6d6b {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.778072] env[61006]: INFO nova.compute.manager [req-d06a084a-7029-4ba7-8629-04ca9b8d7cac req-5291f294-25b5-4f19-be66-0855d9f95887 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Neutron deleted interface 3ddc0fab-3451-455f-9b83-3b55726b6d6b; detaching it from the instance and deleting it from the info cache [ 1022.778251] env[61006]: DEBUG nova.network.neutron [req-d06a084a-7029-4ba7-8629-04ca9b8d7cac req-5291f294-25b5-4f19-be66-0855d9f95887 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.801583] env[61006]: DEBUG oslo_concurrency.lockutils [None req-722ab5f7-b88c-4b54-ba91-e4ff7912aafd tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "aedaa5d6-e0f2-492c-a14b-3254863e1f06" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.149s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.190604] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "0111452e-1b4f-499c-932d-f31364d1a14c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.190971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.191225] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.191434] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.191613] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.193768] env[61006]: INFO nova.compute.manager [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Terminating instance [ 1023.195596] env[61006]: DEBUG nova.compute.manager [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.195793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.196635] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c7d4bb-6d63-441b-9c7b-52161f32b126 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.204908] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.205153] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6741b16-f812-4785-a1ed-a7afdac9dfb7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.211831] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1023.211831] env[61006]: value = "task-1337522" [ 1023.211831] env[61006]: _type = "Task" [ 1023.211831] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.219882] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.254056] env[61006]: DEBUG nova.network.neutron [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.283298] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c17e7da-58d6-4457-abd9-b8f6d748534e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.296861] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143cbf72-b7d2-4589-8d9c-aca8765f8c34 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.325058] env[61006]: DEBUG nova.compute.manager [req-d06a084a-7029-4ba7-8629-04ca9b8d7cac req-5291f294-25b5-4f19-be66-0855d9f95887 service nova] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Detach interface failed, port_id=3ddc0fab-3451-455f-9b83-3b55726b6d6b, reason: Instance 9486ebfc-4717-4c97-8b44-ac922737bc78 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1023.410023] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d6e6d0-a834-43a2-b146-c27477790361 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.416841] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050db9b9-70de-455d-b625-1cca593b5987 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.445756] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049b1e45-e288-4967-aa72-b84c2f45dc8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.453251] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4773aed6-5c94-4237-a93b-89feb56a4a39 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.465930] env[61006]: DEBUG nova.compute.provider_tree [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.685557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-5fc791ca-43a0-4f0a-aa47-c64b81fdd47a tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.230s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.721892] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337522, 'name': PowerOffVM_Task, 'duration_secs': 0.202502} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.722176] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.722359] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.722606] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e563849-e0d3-4d07-b52f-40007ba999fa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.756315] env[61006]: INFO nova.compute.manager [-] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Took 1.26 seconds to deallocate network for instance. [ 1023.781329] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.781567] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.781853] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleting the datastore file [datastore2] 0111452e-1b4f-499c-932d-f31364d1a14c {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.782195] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d90421a5-fee7-413f-a862-ebc7f80aa6d8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.788288] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for the task: (returnval){ [ 1023.788288] env[61006]: value = "task-1337524" [ 1023.788288] env[61006]: _type = "Task" [ 1023.788288] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.796137] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.968747] env[61006]: DEBUG nova.scheduler.client.report [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.265057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.297782] env[61006]: DEBUG oslo_vmware.api [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Task: {'id': task-1337524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144566} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.298101] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.298299] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.298487] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.298666] env[61006]: INFO nova.compute.manager [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1024.298906] env[61006]: DEBUG oslo.service.loopingcall [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.299128] env[61006]: DEBUG nova.compute.manager [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.299223] env[61006]: DEBUG nova.network.neutron [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1024.473683] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.474225] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1024.476958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.212s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.477200] env[61006]: DEBUG nova.objects.instance [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lazy-loading 'resources' on Instance uuid 9486ebfc-4717-4c97-8b44-ac922737bc78 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.702117] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.702312] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.702936] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "66c93148-b91a-4d22-84af-f410c8e10875-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.703146] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.703330] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.705637] env[61006]: INFO nova.compute.manager [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Terminating instance [ 1024.707428] env[61006]: DEBUG nova.compute.manager [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.707627] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.708474] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c967c101-2c7a-4bca-9c02-17a22cf3a41a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.716026] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.716281] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64530ab3-cc3a-442d-878c-8d9d20fd7ba3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.722264] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1024.722264] env[61006]: value = "task-1337525" [ 1024.722264] env[61006]: _type = "Task" [ 1024.722264] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.730120] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337525, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.806868] env[61006]: DEBUG nova.compute.manager [req-84166dae-c014-4042-bfdd-edc4e043c73e req-b131f573-797b-45f9-84ce-c770190bf7ec service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Received event network-vif-deleted-ba96f533-658c-49f4-a9a4-48e129d5c197 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.807081] env[61006]: INFO nova.compute.manager [req-84166dae-c014-4042-bfdd-edc4e043c73e req-b131f573-797b-45f9-84ce-c770190bf7ec service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Neutron deleted interface ba96f533-658c-49f4-a9a4-48e129d5c197; detaching it from the instance and deleting it from the info cache [ 1024.807263] env[61006]: DEBUG nova.network.neutron [req-84166dae-c014-4042-bfdd-edc4e043c73e req-b131f573-797b-45f9-84ce-c770190bf7ec service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.980509] env[61006]: DEBUG nova.compute.utils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.984698] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.984868] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.034280] env[61006]: DEBUG nova.policy [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '85ee4670886d4c8c955ed8adc329132a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931103a837fa4b2eb237dd4715ee0713', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.038076] env[61006]: DEBUG nova.network.neutron [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.094472] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8e480e-e78f-474d-9516-a35b183b895d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.101540] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7719838b-2b78-4bcc-93c4-e5a4af395a57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.130685] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bb980c-69b7-48da-a64d-b8c53f071067 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.138194] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9825796-8fa8-4aeb-b6b4-ac3896a6417a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.151373] env[61006]: DEBUG nova.compute.provider_tree [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.231919] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337525, 'name': PowerOffVM_Task, 'duration_secs': 0.190787} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.232182] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.232416] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.232693] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f00687a7-c2a2-495f-b67c-e361bfdcaa64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.271705] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Successfully created port: 66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.288386] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.288602] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.288783] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.289056] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad2e2454-4170-4026-a0c7-318443f04a9d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.295094] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1025.295094] env[61006]: value = "task-1337527" [ 1025.295094] env[61006]: _type = "Task" [ 1025.295094] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.302500] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.310130] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92b9f920-72ab-4727-bdd8-fc01dc43d267 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.317265] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6977f84b-30a7-4638-b918-ea7366d78f7c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.342285] env[61006]: DEBUG nova.compute.manager [req-84166dae-c014-4042-bfdd-edc4e043c73e req-b131f573-797b-45f9-84ce-c770190bf7ec service nova] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Detach interface failed, port_id=ba96f533-658c-49f4-a9a4-48e129d5c197, reason: Instance 0111452e-1b4f-499c-932d-f31364d1a14c could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1025.485758] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.543041] env[61006]: INFO nova.compute.manager [-] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Took 1.24 seconds to deallocate network for instance. [ 1025.654257] env[61006]: DEBUG nova.scheduler.client.report [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.805237] env[61006]: DEBUG oslo_vmware.api [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131611} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.805501] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.805679] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.805861] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.806052] env[61006]: INFO nova.compute.manager [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1025.806301] env[61006]: DEBUG oslo.service.loopingcall [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.806499] env[61006]: DEBUG nova.compute.manager [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.806594] env[61006]: DEBUG nova.network.neutron [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.051838] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.159402] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.162158] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.110s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.162454] env[61006]: DEBUG nova.objects.instance [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lazy-loading 'resources' on Instance uuid 0111452e-1b4f-499c-932d-f31364d1a14c {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.191698] env[61006]: INFO nova.scheduler.client.report [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Deleted allocations for instance 9486ebfc-4717-4c97-8b44-ac922737bc78 [ 1026.429075] env[61006]: DEBUG nova.compute.manager [req-4e6880ca-37d4-4068-924f-5b2e77eb212e req-dbd51fa2-8258-4170-b665-8a41d8bfdd18 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Received event network-vif-deleted-9230a732-200c-4084-8c6c-a5892e9a50ba {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.429349] env[61006]: INFO nova.compute.manager [req-4e6880ca-37d4-4068-924f-5b2e77eb212e req-dbd51fa2-8258-4170-b665-8a41d8bfdd18 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Neutron deleted interface 9230a732-200c-4084-8c6c-a5892e9a50ba; detaching it from the instance and deleting it from the info cache [ 1026.429572] env[61006]: DEBUG nova.network.neutron [req-4e6880ca-37d4-4068-924f-5b2e77eb212e req-dbd51fa2-8258-4170-b665-8a41d8bfdd18 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.495272] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.521110] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.521364] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.521524] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.521708] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.521853] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.521999] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.523168] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.523358] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.523535] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.523706] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.523884] env[61006]: DEBUG nova.virt.hardware [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.524762] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7116330-4345-4ae9-9bcc-235bd8e60200 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.532740] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d542f266-a95a-4038-9007-19e01ec45bbf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.701094] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a75580e2-1c47-4537-8ea7-11733c6c4c1b tempest-InstanceActionsV221TestJSON-1404016290 tempest-InstanceActionsV221TestJSON-1404016290-project-member] Lock "9486ebfc-4717-4c97-8b44-ac922737bc78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.317s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.766534] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7214a84a-aef5-467a-9e81-0220402d36d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.774131] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6419c492-ae1b-4ddb-9e18-24361a0dddfa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.805418] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb5bb91-6704-431d-9ef2-6a6fe91183d4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.812326] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51aaca9d-ba57-4a83-a084-3bce999de1a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.825090] env[61006]: DEBUG nova.compute.provider_tree [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.834969] env[61006]: DEBUG nova.compute.manager [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Received event network-vif-plugged-66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.835180] env[61006]: DEBUG oslo_concurrency.lockutils [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] Acquiring lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.835285] env[61006]: DEBUG oslo_concurrency.lockutils [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.835415] env[61006]: DEBUG oslo_concurrency.lockutils [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.835586] env[61006]: DEBUG nova.compute.manager [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] No waiting events found dispatching network-vif-plugged-66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1026.835753] env[61006]: WARNING nova.compute.manager [req-06fb8c10-b3b9-431b-a556-6fc906e3319a req-2662d1d5-5066-4caf-8478-a1e614ecda69 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Received unexpected event network-vif-plugged-66dcf158-e221-4c7d-89c6-6384af6d5b21 for instance with vm_state building and task_state spawning. [ 1026.883409] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Successfully updated port: 66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.906166] env[61006]: DEBUG nova.network.neutron [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.932199] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51da4844-be11-467c-9ad4-90991a8d6745 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.942373] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e61f681-35d7-468b-b622-8ab1c5ba2959 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.966013] env[61006]: DEBUG nova.compute.manager [req-4e6880ca-37d4-4068-924f-5b2e77eb212e req-dbd51fa2-8258-4170-b665-8a41d8bfdd18 service nova] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Detach interface failed, port_id=9230a732-200c-4084-8c6c-a5892e9a50ba, reason: Instance 66c93148-b91a-4d22-84af-f410c8e10875 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.328329] env[61006]: DEBUG nova.scheduler.client.report [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.387931] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.388152] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.388279] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.408170] env[61006]: INFO nova.compute.manager [-] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Took 1.60 seconds to deallocate network for instance. [ 1027.833547] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.857764] env[61006]: INFO nova.scheduler.client.report [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Deleted allocations for instance 0111452e-1b4f-499c-932d-f31364d1a14c [ 1027.915391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.915649] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.915869] env[61006]: DEBUG nova.objects.instance [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'resources' on Instance uuid 66c93148-b91a-4d22-84af-f410c8e10875 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.924212] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.045373] env[61006]: DEBUG nova.network.neutron [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Updating instance_info_cache with network_info: [{"id": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "address": "fa:16:3e:d8:11:0b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66dcf158-e2", "ovs_interfaceid": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.365682] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3caef2f-67c0-4b4d-bd73-9d86e45d1308 tempest-ServerRescueNegativeTestJSON-1201311198 tempest-ServerRescueNegativeTestJSON-1201311198-project-member] Lock "0111452e-1b4f-499c-932d-f31364d1a14c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.175s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.518831] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c7ca99-627d-444f-8eeb-948fb1a231ab {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.527243] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f00c25-2b50-4a76-b4b7-39b368eb2dda {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.561639] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.561639] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Instance network_info: |[{"id": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "address": "fa:16:3e:d8:11:0b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66dcf158-e2", "ovs_interfaceid": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1028.562517] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:11:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ecc4615-18f0-4324-8e16-5e5d513325e2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66dcf158-e221-4c7d-89c6-6384af6d5b21', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.570702] env[61006]: DEBUG oslo.service.loopingcall [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.571409] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126e7866-04ee-41d6-a485-596507edb362 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.574149] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.574380] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a6fc8ee-13b1-452e-972c-d3215d5c2bfe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.598828] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a07f954-c9f0-4e7b-9c9c-4c5b34281bd8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.605413] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.605413] env[61006]: value = "task-1337528" [ 1028.605413] env[61006]: _type = "Task" [ 1028.605413] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.618639] env[61006]: DEBUG nova.compute.provider_tree [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.628178] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337528, 'name': CreateVM_Task} progress is 15%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.873528] env[61006]: DEBUG nova.compute.manager [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Received event network-changed-66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.873768] env[61006]: DEBUG nova.compute.manager [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Refreshing instance network info cache due to event network-changed-66dcf158-e221-4c7d-89c6-6384af6d5b21. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.873922] env[61006]: DEBUG oslo_concurrency.lockutils [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] Acquiring lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.875285] env[61006]: DEBUG oslo_concurrency.lockutils [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] Acquired lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.875363] env[61006]: DEBUG nova.network.neutron [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Refreshing network info cache for port 66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.119833] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337528, 'name': CreateVM_Task, 'duration_secs': 0.319171} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.120029] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1029.120702] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.120914] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.121263] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.121532] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79b83a31-c32c-4c02-a3bc-b44f85bbc3c8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.126818] env[61006]: DEBUG nova.scheduler.client.report [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.130640] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1029.130640] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522cdbbd-b02d-a425-cfc8-932dfd5a8666" [ 1029.130640] env[61006]: _type = "Task" [ 1029.130640] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.139471] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522cdbbd-b02d-a425-cfc8-932dfd5a8666, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.634958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.646637] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522cdbbd-b02d-a425-cfc8-932dfd5a8666, 'name': SearchDatastore_Task, 'duration_secs': 0.011996} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.647456] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.647720] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.647967] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.648167] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.648318] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.648589] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5541383-5a37-42cc-9035-0454e63dd934 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.658039] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.658039] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.658039] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c66d6013-bd3a-4651-8248-0f866144e66e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.663494] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1029.663494] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cf9f5f-e03d-396c-9ad9-fedb03f51c4b" [ 1029.663494] env[61006]: _type = "Task" [ 1029.663494] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.667516] env[61006]: INFO nova.scheduler.client.report [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted allocations for instance 66c93148-b91a-4d22-84af-f410c8e10875 [ 1029.674276] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52cf9f5f-e03d-396c-9ad9-fedb03f51c4b, 'name': SearchDatastore_Task, 'duration_secs': 0.008257} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.675295] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92cd1e6f-97cc-4311-b007-3274ab50beda {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.682661] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1029.682661] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522226d1-5758-a8d4-4052-c6f573e02780" [ 1029.682661] env[61006]: _type = "Task" [ 1029.682661] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.692280] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522226d1-5758-a8d4-4052-c6f573e02780, 'name': SearchDatastore_Task, 'duration_secs': 0.009089} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.692557] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.692846] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d7608c4b-a8ec-4e56-bcc3-165aff76f649/d7608c4b-a8ec-4e56-bcc3-165aff76f649.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.693125] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90078984-6ede-48bb-8155-c0ddc4ee991b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.699785] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1029.699785] env[61006]: value = "task-1337529" [ 1029.699785] env[61006]: _type = "Task" [ 1029.699785] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.707857] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.801295] env[61006]: DEBUG nova.network.neutron [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Updated VIF entry in instance network info cache for port 66dcf158-e221-4c7d-89c6-6384af6d5b21. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.801675] env[61006]: DEBUG nova.network.neutron [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Updating instance_info_cache with network_info: [{"id": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "address": "fa:16:3e:d8:11:0b", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66dcf158-e2", "ovs_interfaceid": "66dcf158-e221-4c7d-89c6-6384af6d5b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.178477] env[61006]: DEBUG oslo_concurrency.lockutils [None req-fa3672a5-01d0-4ba2-924e-72628b4b5901 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "66c93148-b91a-4d22-84af-f410c8e10875" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.476s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.209677] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474584} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.210397] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] d7608c4b-a8ec-4e56-bcc3-165aff76f649/d7608c4b-a8ec-4e56-bcc3-165aff76f649.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.210397] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.210571] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc445ddf-70b6-42fd-9876-4967d66379a8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.216772] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1030.216772] env[61006]: value = "task-1337530" [ 1030.216772] env[61006]: _type = "Task" [ 1030.216772] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.225981] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337530, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.304208] env[61006]: DEBUG oslo_concurrency.lockutils [req-94e36461-4eff-4cfe-b645-cfb81244ad9e req-79796653-9142-4adb-85a8-173d9f905cd4 service nova] Releasing lock "refresh_cache-d7608c4b-a8ec-4e56-bcc3-165aff76f649" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.728316] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337530, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062414} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.728754] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.729820] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edf91e0-16cb-422f-a6aa-becdc1711a8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.761023] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] d7608c4b-a8ec-4e56-bcc3-165aff76f649/d7608c4b-a8ec-4e56-bcc3-165aff76f649.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.761023] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e642e51c-6867-4136-8bb9-f01bd959f973 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.784017] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1030.784017] env[61006]: value = "task-1337531" [ 1030.784017] env[61006]: _type = "Task" [ 1030.784017] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.791978] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337531, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.294207] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337531, 'name': ReconfigVM_Task, 'duration_secs': 0.509015} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.295321] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Reconfigured VM instance instance-00000069 to attach disk [datastore2] d7608c4b-a8ec-4e56-bcc3-165aff76f649/d7608c4b-a8ec-4e56-bcc3-165aff76f649.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.296190] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68546605-e5c9-4f0e-b8f7-3854984d3958 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.304409] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1031.304409] env[61006]: value = "task-1337532" [ 1031.304409] env[61006]: _type = "Task" [ 1031.304409] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.313674] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337532, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.815964] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337532, 'name': Rename_Task, 'duration_secs': 0.132626} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.817037] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.817037] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c82399f-ec2e-4c2f-878b-59aeb1838925 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.823322] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1031.823322] env[61006]: value = "task-1337533" [ 1031.823322] env[61006]: _type = "Task" [ 1031.823322] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.831055] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337533, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.008308] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.011023] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.011023] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1032.011023] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1032.078996] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.079255] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.334960] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337533, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.513681] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Skipping network cache update for instance because it is Building. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1032.546138] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.546295] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquired lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.546444] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Forcefully refreshing network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1032.546598] env[61006]: DEBUG nova.objects.instance [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lazy-loading 'info_cache' on Instance uuid 42061ea3-d1d1-4633-bd24-65f7ee302c1f {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.582082] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.834174] env[61006]: DEBUG oslo_vmware.api [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337533, 'name': PowerOnVM_Task, 'duration_secs': 0.546416} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.834481] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.834686] env[61006]: INFO nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Took 6.34 seconds to spawn the instance on the hypervisor. [ 1032.834874] env[61006]: DEBUG nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.835708] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b62f73e-dffe-41ca-80c6-4c659acb9eea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.105235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.105235] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.107215] env[61006]: INFO nova.compute.claims [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.354708] env[61006]: INFO nova.compute.manager [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Took 11.60 seconds to build instance. [ 1033.856543] env[61006]: DEBUG oslo_concurrency.lockutils [None req-57b0e5c8-2b90-4658-a444-64bd7097fbf1 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.114s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.193689] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e30596a-a20d-4ae5-9533-ab0f442fad44 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.202993] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409f60a6-8f84-4e3e-9b97-10936efd0b2b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.233879] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4dfcec-8b56-4be6-a9db-f257f6863ca0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.242916] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdea486-3577-40af-ae40-da434b3e60e2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.256756] env[61006]: DEBUG nova.compute.provider_tree [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1034.272255] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updating instance_info_cache with network_info: [{"id": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "address": "fa:16:3e:3d:5c:fc", "network": {"id": "f81a3264-103b-40fb-945e-fcf7a30dd112", "bridge": "br-int", "label": "tempest-ServersTestJSON-1488699940-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931103a837fa4b2eb237dd4715ee0713", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ecc4615-18f0-4324-8e16-5e5d513325e2", "external-id": "nsx-vlan-transportzone-167", "segmentation_id": 167, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb597d7e4-0d", "ovs_interfaceid": "b597d7e4-0db4-40c4-90a2-f2245d7cdc67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.385272] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.385585] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.385745] env[61006]: DEBUG nova.compute.manager [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.386661] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b3ee8b-cd47-4d50-8998-76f6ae267919 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.393868] env[61006]: DEBUG nova.compute.manager [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61006) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1034.394632] env[61006]: DEBUG nova.objects.instance [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'flavor' on Instance uuid d7608c4b-a8ec-4e56-bcc3-165aff76f649 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.774325] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Releasing lock "refresh_cache-42061ea3-d1d1-4633-bd24-65f7ee302c1f" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.774529] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updated the network info_cache for instance {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1034.774734] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.774911] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.775142] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.775311] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.776342] env[61006]: ERROR nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [req-671fbe62-68c7-487e-8437-9bea07a074d7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3360045e-46ab-4f2d-9377-dd481ab3cd53. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-671fbe62-68c7-487e-8437-9bea07a074d7"}]} [ 1034.776660] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.778724] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.778869] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1034.779043] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.795517] env[61006]: DEBUG nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Refreshing inventories for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1034.810264] env[61006]: DEBUG nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating ProviderTree inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1034.810495] env[61006]: DEBUG nova.compute.provider_tree [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1034.821626] env[61006]: DEBUG nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Refreshing aggregate associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, aggregates: None {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1034.840337] env[61006]: DEBUG nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Refreshing trait associations for resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61006) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1034.899734] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.899972] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59603054-2501-4680-a8c1-d316e4907071 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.909692] env[61006]: DEBUG oslo_vmware.api [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1034.909692] env[61006]: value = "task-1337537" [ 1034.909692] env[61006]: _type = "Task" [ 1034.909692] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.922802] env[61006]: DEBUG oslo_vmware.api [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.927035] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9960a4-2a13-4722-9e32-07a590aca66f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.933858] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c889da-00e7-44b0-b02a-0e29c3ba3c78 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.963184] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19758327-ee8b-4253-aecf-eb4c3880ca55 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.970309] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7e6440-682f-468c-98a9-62dd25081cbd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.985752] env[61006]: DEBUG nova.compute.provider_tree [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.282160] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.421608] env[61006]: DEBUG oslo_vmware.api [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337537, 'name': PowerOffVM_Task, 'duration_secs': 0.248025} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.422231] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.422231] env[61006]: DEBUG nova.compute.manager [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.423117] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3526d1b-6218-4c1c-9271-ab6b80b96348 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.525591] env[61006]: DEBUG nova.scheduler.client.report [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 128 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1035.525854] env[61006]: DEBUG nova.compute.provider_tree [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 128 to 129 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1035.526187] env[61006]: DEBUG nova.compute.provider_tree [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1035.937814] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c9cd356c-e36c-4494-9195-d22f10be0918 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.031978] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.926s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.032284] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.035321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.753s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.035411] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.035569] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1036.036653] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95b855e-1e88-4575-a531-025969915e8a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.045852] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e4d873-53ad-4297-8c35-d0bd54dbc2b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.059959] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37f923c-9e18-4c2e-952c-e02102b68707 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.066831] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9204ef4d-cb3f-4b6a-930e-7e2d99b168ff {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.098395] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180073MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1036.098560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.098757] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.537612] env[61006]: DEBUG nova.compute.utils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.538952] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.539137] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.603010] env[61006]: DEBUG nova.policy [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8865ab804d34ccba8909c60cdb51adb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f60c05599bb7457f9bd7a3d11daf9ab3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.859583] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Successfully created port: 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.986599] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.986833] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.987052] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.987242] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.987415] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.991732] env[61006]: INFO nova.compute.manager [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Terminating instance [ 1036.994614] env[61006]: DEBUG nova.compute.manager [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1036.994818] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.995678] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f29d737-68a3-4615-8bf4-80e05e8f765a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.003601] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.003832] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-555c9e86-79b2-4410-81d5-171a680662ad {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.042753] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.062880] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.064031] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.064031] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore2] d7608c4b-a8ec-4e56-bcc3-165aff76f649 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.064031] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d1fca3d-2d82-4273-a96b-f924e8ad5747 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.070255] env[61006]: DEBUG oslo_vmware.api [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1037.070255] env[61006]: value = "task-1337540" [ 1037.070255] env[61006]: _type = "Task" [ 1037.070255] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.078421] env[61006]: DEBUG oslo_vmware.api [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.128513] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.128669] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 02408ca9-c580-444a-9608-a752146ca499 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.128794] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance c8d841a2-218a-48d7-8716-e47c29798b00 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.128914] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 5c1c8651-cb82-49c7-9e0d-fe3960fc1633 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.129043] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance d7608c4b-a8ec-4e56-bcc3-165aff76f649 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.129159] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1037.129341] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1037.129502] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1037.219811] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9617092-5290-48fb-9441-b44cec20bf8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.227403] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48da3c9e-4858-4713-8557-b1fc5b32b7cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.257341] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38a91a7-dbab-4f37-bf4b-7eca0f94f27a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.263899] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47cae36-eba8-4c84-b0ca-f13645d93f8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.276421] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.580983] env[61006]: DEBUG oslo_vmware.api [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151427} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.581945] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1037.582176] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1037.582373] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1037.582560] env[61006]: INFO nova.compute.manager [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1037.582845] env[61006]: DEBUG oslo.service.loopingcall [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.583289] env[61006]: DEBUG nova.compute.manager [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1037.583382] env[61006]: DEBUG nova.network.neutron [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.807566] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 129 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1037.807792] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 129 to 130 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1037.807941] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.864688] env[61006]: DEBUG nova.compute.manager [req-71fd48c8-816b-4947-af98-fa0510e2c23f req-76258056-3a05-44f1-acc2-bd2664bf18d0 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Received event network-vif-deleted-66dcf158-e221-4c7d-89c6-6384af6d5b21 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.864951] env[61006]: INFO nova.compute.manager [req-71fd48c8-816b-4947-af98-fa0510e2c23f req-76258056-3a05-44f1-acc2-bd2664bf18d0 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Neutron deleted interface 66dcf158-e221-4c7d-89c6-6384af6d5b21; detaching it from the instance and deleting it from the info cache [ 1037.865158] env[61006]: DEBUG nova.network.neutron [req-71fd48c8-816b-4947-af98-fa0510e2c23f req-76258056-3a05-44f1-acc2-bd2664bf18d0 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.052800] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.079197] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.079457] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.079620] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.079805] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.079956] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.080126] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.080337] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.080497] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.080663] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.080894] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.081125] env[61006]: DEBUG nova.virt.hardware [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.082148] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d67352-8ba0-440d-9ef5-a1cdc93acc42 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.090853] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24108d0a-2b48-4819-8092-bdbe40149905 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.251996] env[61006]: DEBUG nova.compute.manager [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.251996] env[61006]: DEBUG oslo_concurrency.lockutils [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.251996] env[61006]: DEBUG oslo_concurrency.lockutils [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.251996] env[61006]: DEBUG oslo_concurrency.lockutils [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.251996] env[61006]: DEBUG nova.compute.manager [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] No waiting events found dispatching network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.251996] env[61006]: WARNING nova.compute.manager [req-6165a67b-09e6-4fc7-8b32-2ea6beddcb34 req-96446676-a839-42d1-8279-bebfbebf6b7a service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received unexpected event network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d for instance with vm_state building and task_state spawning. [ 1038.313540] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1038.313540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.215s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.334605] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Successfully updated port: 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.342870] env[61006]: DEBUG nova.network.neutron [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.368084] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3cffd35-4d21-411b-87d5-0e3e6c42bc4b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.379261] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04805d88-02fe-4afe-8443-49529073066f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.408236] env[61006]: DEBUG nova.compute.manager [req-71fd48c8-816b-4947-af98-fa0510e2c23f req-76258056-3a05-44f1-acc2-bd2664bf18d0 service nova] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Detach interface failed, port_id=66dcf158-e221-4c7d-89c6-6384af6d5b21, reason: Instance d7608c4b-a8ec-4e56-bcc3-165aff76f649 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1038.837443] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.837727] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.837727] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.844927] env[61006]: INFO nova.compute.manager [-] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Took 1.26 seconds to deallocate network for instance. [ 1039.351642] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.351920] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.352229] env[61006]: DEBUG nova.objects.instance [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid d7608c4b-a8ec-4e56-bcc3-165aff76f649 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.367919] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1039.499535] env[61006]: DEBUG nova.network.neutron [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.937759] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cd81f6-e3eb-47ac-ab0d-427af6ec8352 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.945280] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd0f830-8777-4994-9f53-527b82b21cf7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.974198] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6290bd-d7eb-4764-913a-cce8e473cfae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.980763] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7981f1d4-95ae-443d-8c96-e5e96b957e58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.994117] env[61006]: DEBUG nova.compute.provider_tree [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.002168] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.002802] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance network_info: |[{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1040.002802] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:87:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b4cb734-5640-4e37-98ac-a329e9c9562d', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.010350] env[61006]: DEBUG oslo.service.loopingcall [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.011163] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.011373] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e044c58-4bc1-4ca0-a5e0-b543e4b036ae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.029822] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.029822] env[61006]: value = "task-1337543" [ 1040.029822] env[61006]: _type = "Task" [ 1040.029822] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.036679] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337543, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.275783] env[61006]: DEBUG nova.compute.manager [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.276025] env[61006]: DEBUG nova.compute.manager [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing instance network info cache due to event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.276203] env[61006]: DEBUG oslo_concurrency.lockutils [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.276356] env[61006]: DEBUG oslo_concurrency.lockutils [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.276507] env[61006]: DEBUG nova.network.neutron [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.496895] env[61006]: DEBUG nova.scheduler.client.report [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.539080] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337543, 'name': CreateVM_Task, 'duration_secs': 0.277936} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.539285] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1040.539878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.540057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.540373] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.540618] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b004d3-0404-478e-87f6-f2cea4ac9cc7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.544706] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1040.544706] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ebf291-eb2a-b75f-f731-99530b4be059" [ 1040.544706] env[61006]: _type = "Task" [ 1040.544706] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.551632] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ebf291-eb2a-b75f-f731-99530b4be059, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.955051] env[61006]: DEBUG nova.network.neutron [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updated VIF entry in instance network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.955462] env[61006]: DEBUG nova.network.neutron [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.001932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.020094] env[61006]: INFO nova.scheduler.client.report [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance d7608c4b-a8ec-4e56-bcc3-165aff76f649 [ 1041.055098] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ebf291-eb2a-b75f-f731-99530b4be059, 'name': SearchDatastore_Task, 'duration_secs': 0.009958} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.055634] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.055634] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.055882] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.056046] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.056236] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.056493] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f22f56df-011d-414e-80b1-2c0376809094 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.064843] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.065032] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1041.065798] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f16f3784-801d-4b10-bf7c-700348069fee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.070967] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1041.070967] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521392f4-7c01-19ea-cda1-7caf99ef874b" [ 1041.070967] env[61006]: _type = "Task" [ 1041.070967] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.078273] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521392f4-7c01-19ea-cda1-7caf99ef874b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.458708] env[61006]: DEBUG oslo_concurrency.lockutils [req-db3473ab-2fc7-4910-bee8-ba3589604f30 req-3af09c9e-89f3-4e7f-b218-881e6c626b21 service nova] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.528125] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e3989202-6022-4384-8c68-62bd40113646 tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "d7608c4b-a8ec-4e56-bcc3-165aff76f649" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.541s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.581121] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]521392f4-7c01-19ea-cda1-7caf99ef874b, 'name': SearchDatastore_Task, 'duration_secs': 0.008024} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.581904] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcbfed00-c26b-4f3d-8335-590ddba24509 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.586761] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1041.586761] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e22df0-f980-4765-9875-2582e109212e" [ 1041.586761] env[61006]: _type = "Task" [ 1041.586761] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.596391] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e22df0-f980-4765-9875-2582e109212e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.098543] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52e22df0-f980-4765-9875-2582e109212e, 'name': SearchDatastore_Task, 'duration_secs': 0.009796} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.098874] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.099257] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1042.099553] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea7de374-575a-45da-a86e-1203662dd9be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.106722] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1042.106722] env[61006]: value = "task-1337545" [ 1042.106722] env[61006]: _type = "Task" [ 1042.106722] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.115181] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.619616] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337545, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.835966] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.836334] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.836623] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.836840] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.837130] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.839424] env[61006]: INFO nova.compute.manager [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Terminating instance [ 1042.841296] env[61006]: DEBUG nova.compute.manager [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.841496] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.842327] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3551920b-63df-4f13-b5bf-e8d23217c606 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.850342] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.850639] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6927ac00-5691-49ef-acde-8f4fefb2c6ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.856569] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1042.856569] env[61006]: value = "task-1337546" [ 1042.856569] env[61006]: _type = "Task" [ 1042.856569] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.864484] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337546, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.116905] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561853} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.117249] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1043.117392] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.117706] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c6c02c6-3d8a-42d0-bdfa-4326f2bc48c2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.124479] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1043.124479] env[61006]: value = "task-1337547" [ 1043.124479] env[61006]: _type = "Task" [ 1043.124479] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.133373] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.367036] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337546, 'name': PowerOffVM_Task, 'duration_secs': 0.233184} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.367036] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.367165] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.367399] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8980845a-0736-402c-8028-ae4a8e0235b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.425741] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.426021] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.426213] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleting the datastore file [datastore2] 42061ea3-d1d1-4633-bd24-65f7ee302c1f {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.426455] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52d3f7ea-e908-4f9b-a804-4fc87c5aeeec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.432804] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for the task: (returnval){ [ 1043.432804] env[61006]: value = "task-1337549" [ 1043.432804] env[61006]: _type = "Task" [ 1043.432804] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.440555] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.634069] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131587} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.634374] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.635153] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450e5239-93b7-4d1c-8bf7-77a3944adf94 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.656038] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.656285] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1a25e30-76f4-437d-8812-c8cc4f2dcf00 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.676706] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1043.676706] env[61006]: value = "task-1337550" [ 1043.676706] env[61006]: _type = "Task" [ 1043.676706] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.686979] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.942065] env[61006]: DEBUG oslo_vmware.api [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Task: {'id': task-1337549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142596} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.942395] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.942609] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.942793] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.943009] env[61006]: INFO nova.compute.manager [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1043.943332] env[61006]: DEBUG oslo.service.loopingcall [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.943558] env[61006]: DEBUG nova.compute.manager [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.943654] env[61006]: DEBUG nova.network.neutron [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.176996] env[61006]: DEBUG nova.compute.manager [req-65086d64-1c84-4722-95eb-51d0db725fcf req-634c6db6-c60d-4a22-8f17-2a358a832640 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Received event network-vif-deleted-b597d7e4-0db4-40c4-90a2-f2245d7cdc67 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.177238] env[61006]: INFO nova.compute.manager [req-65086d64-1c84-4722-95eb-51d0db725fcf req-634c6db6-c60d-4a22-8f17-2a358a832640 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Neutron deleted interface b597d7e4-0db4-40c4-90a2-f2245d7cdc67; detaching it from the instance and deleting it from the info cache [ 1044.177387] env[61006]: DEBUG nova.network.neutron [req-65086d64-1c84-4722-95eb-51d0db725fcf req-634c6db6-c60d-4a22-8f17-2a358a832640 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.188220] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337550, 'name': ReconfigVM_Task, 'duration_secs': 0.417116} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.188589] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.189617] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2ecfd48-11b1-4113-beaa-dacdd1b30cd5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.195637] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1044.195637] env[61006]: value = "task-1337552" [ 1044.195637] env[61006]: _type = "Task" [ 1044.195637] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.203421] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337552, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.651584] env[61006]: DEBUG nova.network.neutron [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.680428] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f940da4c-b07f-44f5-b48a-bba87ccc2318 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.690020] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e29fa1-3848-421d-bec7-e3fb8cf0c910 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.705569] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.705770] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.712794] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337552, 'name': Rename_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.719043] env[61006]: DEBUG nova.compute.manager [req-65086d64-1c84-4722-95eb-51d0db725fcf req-634c6db6-c60d-4a22-8f17-2a358a832640 service nova] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Detach interface failed, port_id=b597d7e4-0db4-40c4-90a2-f2245d7cdc67, reason: Instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.154725] env[61006]: INFO nova.compute.manager [-] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Took 1.21 seconds to deallocate network for instance. [ 1045.211123] env[61006]: DEBUG nova.compute.utils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.212957] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337552, 'name': Rename_Task, 'duration_secs': 0.968723} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.213199] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.213468] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbff6276-ac33-47fd-b9f6-e56b6334789b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.220338] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1045.220338] env[61006]: value = "task-1337553" [ 1045.220338] env[61006]: _type = "Task" [ 1045.220338] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.234179] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.661881] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.662186] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.662433] env[61006]: DEBUG nova.objects.instance [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lazy-loading 'resources' on Instance uuid 42061ea3-d1d1-4633-bd24-65f7ee302c1f {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.714536] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.730093] env[61006]: DEBUG oslo_vmware.api [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337553, 'name': PowerOnVM_Task, 'duration_secs': 0.43525} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.730355] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.730552] env[61006]: INFO nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Took 7.68 seconds to spawn the instance on the hypervisor. [ 1045.730727] env[61006]: DEBUG nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.731506] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49525f4d-b171-43c9-b7f7-ca2f4f83fe6d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.234782] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ac7e53-1169-4fab-9f08-4856d5581579 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.246076] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abac4af6-257f-4c18-a98b-7615907be82c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.251138] env[61006]: INFO nova.compute.manager [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Took 13.16 seconds to build instance. [ 1046.278655] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12687f6-51b2-43c0-b869-3ab38dae972b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.288961] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c64fce-0e22-4b72-aa51-4451c67cae0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.304166] env[61006]: DEBUG nova.compute.provider_tree [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1046.754011] env[61006]: DEBUG oslo_concurrency.lockutils [None req-28a16e20-27e9-476c-885e-20647db19b48 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.675s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.784116] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.784400] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.784639] env[61006]: INFO nova.compute.manager [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Attaching volume 37a60d5e-9349-4506-b970-81e4f379624f to /dev/sdb [ 1046.815162] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c71618f-1ce2-481a-9677-1e22d29d3301 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.822572] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197b4326-abe2-4f53-8ac2-4a9f038f20f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.836560] env[61006]: DEBUG nova.virt.block_device [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating existing volume attachment record: db80f0af-802d-4e58-9516-e68570786693 {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1046.839211] env[61006]: DEBUG nova.scheduler.client.report [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updated inventory for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1046.839440] env[61006]: DEBUG nova.compute.provider_tree [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating resource provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 generation from 130 to 131 during operation: update_inventory {{(pid=61006) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1046.839614] env[61006]: DEBUG nova.compute.provider_tree [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Updating inventory in ProviderTree for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1047.343929] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.373453] env[61006]: INFO nova.scheduler.client.report [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Deleted allocations for instance 42061ea3-d1d1-4633-bd24-65f7ee302c1f [ 1047.593516] env[61006]: DEBUG nova.compute.manager [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.593922] env[61006]: DEBUG nova.compute.manager [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing instance network info cache due to event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.593922] env[61006]: DEBUG oslo_concurrency.lockutils [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.595211] env[61006]: DEBUG oslo_concurrency.lockutils [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.595211] env[61006]: DEBUG nova.network.neutron [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.883059] env[61006]: DEBUG oslo_concurrency.lockutils [None req-f6564a9d-dbfa-410c-9d50-fdef7314a51a tempest-ServersTestJSON-1445808535 tempest-ServersTestJSON-1445808535-project-member] Lock "42061ea3-d1d1-4633-bd24-65f7ee302c1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.046s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.321039] env[61006]: DEBUG nova.network.neutron [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updated VIF entry in instance network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.321424] env[61006]: DEBUG nova.network.neutron [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.824650] env[61006]: DEBUG oslo_concurrency.lockutils [req-f6076f66-4d16-4132-a2e1-6fc78027ae3f req-11c0efd7-22c9-4507-991a-5253093eb188 service nova] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.116086] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.116418] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.385926] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1051.386190] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285431', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'name': 'volume-37a60d5e-9349-4506-b970-81e4f379624f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8d841a2-218a-48d7-8716-e47c29798b00', 'attached_at': '', 'detached_at': '', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'serial': '37a60d5e-9349-4506-b970-81e4f379624f'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1051.387091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c10d27d-b9eb-45ec-a4cd-f179d59beeca {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.404823] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c223943f-4231-48c6-b32b-691adb2fabed {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.429123] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] volume-37a60d5e-9349-4506-b970-81e4f379624f/volume-37a60d5e-9349-4506-b970-81e4f379624f.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.429496] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31c9eeda-0471-49af-a281-2e405ec2a395 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.446902] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1051.446902] env[61006]: value = "task-1337560" [ 1051.446902] env[61006]: _type = "Task" [ 1051.446902] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.454478] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337560, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.618560] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.955912] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337560, 'name': ReconfigVM_Task, 'duration_secs': 0.457008} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.956215] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to attach disk [datastore2] volume-37a60d5e-9349-4506-b970-81e4f379624f/volume-37a60d5e-9349-4506-b970-81e4f379624f.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.960835] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-755c8a77-4b4f-463b-b260-1aad6d22c8e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.974884] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1051.974884] env[61006]: value = "task-1337561" [ 1051.974884] env[61006]: _type = "Task" [ 1051.974884] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.982371] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337561, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.139470] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.139776] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.141352] env[61006]: INFO nova.compute.claims [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.484660] env[61006]: DEBUG oslo_vmware.api [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337561, 'name': ReconfigVM_Task, 'duration_secs': 0.128935} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.484959] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285431', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'name': 'volume-37a60d5e-9349-4506-b970-81e4f379624f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8d841a2-218a-48d7-8716-e47c29798b00', 'attached_at': '', 'detached_at': '', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'serial': '37a60d5e-9349-4506-b970-81e4f379624f'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1053.226663] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4ba8dd-b493-4e24-bd72-ca71dc3aea69 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.235290] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d735cd-a941-42f9-b84e-7f217117ccbd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.266101] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553819ef-3c07-41a2-808d-52b2779f9337 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.273378] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d4b8e3-dfc8-4d42-867d-ad57b07cb34b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.287557] env[61006]: DEBUG nova.compute.provider_tree [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.523817] env[61006]: DEBUG nova.objects.instance [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'flavor' on Instance uuid c8d841a2-218a-48d7-8716-e47c29798b00 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.791235] env[61006]: DEBUG nova.scheduler.client.report [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.028332] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c4e03a6b-8ad7-4689-9a3a-ef6f2a132990 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.244s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.296532] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.296889] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.802290] env[61006]: DEBUG nova.compute.utils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.803794] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1054.803961] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1054.857299] env[61006]: DEBUG nova.policy [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13d543e04e404e9ba9dc140cb19bbe1e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '440694a84a91482cb8363f0e38562833', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1054.858789] env[61006]: DEBUG nova.compute.manager [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Stashing vm_state: active {{(pid=61006) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1055.153585] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Successfully created port: e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.307333] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1055.380684] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.380958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.812139] env[61006]: INFO nova.virt.block_device [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Booting with volume d3b7bb17-e4e9-4255-a237-a79e5010ad0b at /dev/sda [ 1055.848426] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab7cb7dc-c2db-4f22-af6d-b1f9504401d3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.857347] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370e6a4a-8852-4169-89aa-745566c5d6cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.884199] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f7c4eac-272e-47c1-9c12-4b01e6fe7a20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.885475] env[61006]: INFO nova.compute.claims [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.896013] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a9efe2-389d-474d-a207-6177efeebc99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.920769] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85db651-ea10-4e14-a405-2b041a28767a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.927454] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a430cd42-c7f8-45ad-88f1-130db907cea3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.942027] env[61006]: DEBUG nova.virt.block_device [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating existing volume attachment record: 71eed321-6fcc-453e-81d3-415c67dd3f1d {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1056.391232] env[61006]: INFO nova.compute.resource_tracker [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating resource usage from migration 7ee371a3-8cd6-48db-b4c1-032fe8217aac [ 1056.477833] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e3215f-4338-4bf9-8af5-49b26762f5ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.487091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45817cf-1087-49c7-ac79-d86000f3897a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.519781] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4404a1bf-657a-475d-9bcd-e7af43ebffaa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.523184] env[61006]: DEBUG nova.compute.manager [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Received event network-vif-plugged-e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.523416] env[61006]: DEBUG oslo_concurrency.lockutils [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] Acquiring lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.523627] env[61006]: DEBUG oslo_concurrency.lockutils [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.523798] env[61006]: DEBUG oslo_concurrency.lockutils [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.523963] env[61006]: DEBUG nova.compute.manager [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] No waiting events found dispatching network-vif-plugged-e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1056.524159] env[61006]: WARNING nova.compute.manager [req-84a96c40-ff25-4128-8b3b-649b2e0505dc req-7dffe627-2a0b-4de8-8da6-d28b7dcb86c7 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Received unexpected event network-vif-plugged-e41ece56-1df4-4013-bace-827b00f83fee for instance with vm_state building and task_state block_device_mapping. [ 1056.529928] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.530164] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.530417] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.530626] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.530798] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.533077] env[61006]: INFO nova.compute.manager [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Terminating instance [ 1056.535662] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92deaeef-91eb-4254-a4f0-ba1dc91bfb91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.541085] env[61006]: DEBUG nova.compute.manager [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.541298] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.542401] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d45ca9a-dad2-43c2-9ea1-c3d9dba9b705 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.557678] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.558536] env[61006]: DEBUG nova.compute.provider_tree [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.559789] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8906c2b8-63dc-46dc-ad81-856da8522f82 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.565896] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1056.565896] env[61006]: value = "task-1337562" [ 1056.565896] env[61006]: _type = "Task" [ 1056.565896] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.575300] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.606498] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Successfully updated port: e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.066460] env[61006]: DEBUG nova.scheduler.client.report [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.077771] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337562, 'name': PowerOffVM_Task, 'duration_secs': 0.175444} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.078042] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.078221] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.078474] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95601a2e-1759-42bc-9957-36603595870f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.109467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.109631] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquired lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.109767] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.136980] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.137287] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.137520] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleting the datastore file [datastore2] 5c1c8651-cb82-49c7-9e0d-fe3960fc1633 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.138686] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42ee59f5-e563-49f1-800b-1bf1430e7bf1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.144789] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1057.144789] env[61006]: value = "task-1337564" [ 1057.144789] env[61006]: _type = "Task" [ 1057.144789] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.152680] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337564, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.573116] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.192s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.573489] env[61006]: INFO nova.compute.manager [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Migrating [ 1057.639614] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.655339] env[61006]: DEBUG oslo_vmware.api [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337564, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133724} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.655646] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.655847] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.656371] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.656371] env[61006]: INFO nova.compute.manager [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1057.656501] env[61006]: DEBUG oslo.service.loopingcall [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.656647] env[61006]: DEBUG nova.compute.manager [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.656742] env[61006]: DEBUG nova.network.neutron [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.768143] env[61006]: DEBUG nova.network.neutron [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating instance_info_cache with network_info: [{"id": "e41ece56-1df4-4013-bace-827b00f83fee", "address": "fa:16:3e:bb:c5:7b", "network": {"id": "352a1fb9-fc7f-44e2-a74b-8624e667091b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-539177150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "440694a84a91482cb8363f0e38562833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41ece56-1d", "ovs_interfaceid": "e41ece56-1df4-4013-bace-827b00f83fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.030110] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1058.030679] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.030904] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.031073] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.031291] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.031434] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.031582] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.031802] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.032041] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.032183] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.032363] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.032538] env[61006]: DEBUG nova.virt.hardware [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.033466] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269fad52-57a6-4ab4-83af-3ba5067076b7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.042220] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acea66e-922c-4d15-8999-1dad60bdb7ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.088453] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.088639] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.088818] env[61006]: DEBUG nova.network.neutron [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.091684] env[61006]: DEBUG nova.compute.manager [req-946d4e2f-dacd-42a8-9e09-ed848268ac6e req-36303f10-39b9-473a-8680-aa67b9cef0da service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Received event network-vif-deleted-4f0a0867-f2dd-4b43-a42c-05bce1275bb6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.091937] env[61006]: INFO nova.compute.manager [req-946d4e2f-dacd-42a8-9e09-ed848268ac6e req-36303f10-39b9-473a-8680-aa67b9cef0da service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Neutron deleted interface 4f0a0867-f2dd-4b43-a42c-05bce1275bb6; detaching it from the instance and deleting it from the info cache [ 1058.092190] env[61006]: DEBUG nova.network.neutron [req-946d4e2f-dacd-42a8-9e09-ed848268ac6e req-36303f10-39b9-473a-8680-aa67b9cef0da service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.277016] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Releasing lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.277016] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance network_info: |[{"id": "e41ece56-1df4-4013-bace-827b00f83fee", "address": "fa:16:3e:bb:c5:7b", "network": {"id": "352a1fb9-fc7f-44e2-a74b-8624e667091b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-539177150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "440694a84a91482cb8363f0e38562833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41ece56-1d", "ovs_interfaceid": "e41ece56-1df4-4013-bace-827b00f83fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1058.277016] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:c5:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e41ece56-1df4-4013-bace-827b00f83fee', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.282854] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Creating folder: Project (440694a84a91482cb8363f0e38562833). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1058.283171] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bbdc4b9-9a0d-46fe-863d-06170fd6f7c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.295993] env[61006]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1058.300015] env[61006]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61006) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1058.300015] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Folder already exists: Project (440694a84a91482cb8363f0e38562833). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1058.300015] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Creating folder: Instances. Parent ref: group-v285425. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1058.300015] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09442a70-e5a6-40b5-a879-8bdd56fd4b60 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.306788] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Created folder: Instances in parent group-v285425. [ 1058.307040] env[61006]: DEBUG oslo.service.loopingcall [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.307243] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.307448] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3cd21713-50bc-4a82-9761-168cb1a1b082 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.326142] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1058.326142] env[61006]: value = "task-1337567" [ 1058.326142] env[61006]: _type = "Task" [ 1058.326142] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.333095] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337567, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.548421] env[61006]: DEBUG nova.compute.manager [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Received event network-changed-e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.548633] env[61006]: DEBUG nova.compute.manager [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Refreshing instance network info cache due to event network-changed-e41ece56-1df4-4013-bace-827b00f83fee. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.548832] env[61006]: DEBUG oslo_concurrency.lockutils [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] Acquiring lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.549009] env[61006]: DEBUG oslo_concurrency.lockutils [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] Acquired lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.549385] env[61006]: DEBUG nova.network.neutron [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Refreshing network info cache for port e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1058.561959] env[61006]: DEBUG nova.network.neutron [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.596858] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a08b447-f5ea-43d0-aeb6-56c689c6bb09 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.608211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61be7c3-1313-4993-979b-273ebceea6a3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.635951] env[61006]: DEBUG nova.compute.manager [req-946d4e2f-dacd-42a8-9e09-ed848268ac6e req-36303f10-39b9-473a-8680-aa67b9cef0da service nova] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Detach interface failed, port_id=4f0a0867-f2dd-4b43-a42c-05bce1275bb6, reason: Instance 5c1c8651-cb82-49c7-9e0d-fe3960fc1633 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1058.803864] env[61006]: DEBUG nova.network.neutron [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.836116] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337567, 'name': CreateVM_Task, 'duration_secs': 0.296524} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.836281] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.836941] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285429', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'name': 'volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '059d8cea-e1d9-4e27-828f-e0256c83b1f1', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'serial': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b'}, 'guest_format': None, 'boot_index': 0, 'mount_device': '/dev/sda', 'disk_bus': None, 'attachment_id': '71eed321-6fcc-453e-81d3-415c67dd3f1d', 'device_type': None, 'delete_on_termination': True, 'volume_type': None}], 'swap': None} {{(pid=61006) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1058.837174] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Root volume attach. Driver type: vmdk {{(pid=61006) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1058.837926] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9394577d-3399-4f2e-9f02-b05fff13f0cb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.845059] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae7b409-6ea2-4559-83e5-7028db067ca8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.851031] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5ee752-cf12-4b8b-b419-5604b14e5c8e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.856499] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-110672dc-530a-4cbe-8588-78514162e829 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.863963] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1058.863963] env[61006]: value = "task-1337568" [ 1058.863963] env[61006]: _type = "Task" [ 1058.863963] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.872545] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337568, 'name': RelocateVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.064697] env[61006]: INFO nova.compute.manager [-] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Took 1.41 seconds to deallocate network for instance. [ 1059.242557] env[61006]: DEBUG nova.network.neutron [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updated VIF entry in instance network info cache for port e41ece56-1df4-4013-bace-827b00f83fee. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.242935] env[61006]: DEBUG nova.network.neutron [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating instance_info_cache with network_info: [{"id": "e41ece56-1df4-4013-bace-827b00f83fee", "address": "fa:16:3e:bb:c5:7b", "network": {"id": "352a1fb9-fc7f-44e2-a74b-8624e667091b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-539177150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "440694a84a91482cb8363f0e38562833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41ece56-1d", "ovs_interfaceid": "e41ece56-1df4-4013-bace-827b00f83fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.306479] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.373626] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337568, 'name': RelocateVM_Task, 'duration_secs': 0.347098} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.373893] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1059.374114] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285429', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'name': 'volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '059d8cea-e1d9-4e27-828f-e0256c83b1f1', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'serial': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1059.374894] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83a02d2-92e2-4be7-9528-51c5ec39ab78 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.389833] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d9a823-6a35-45b1-933a-96a024d6b34e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.411897] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b/volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.412143] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-978d7caa-ec2d-471c-8128-d7f7c4a886f0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.430633] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1059.430633] env[61006]: value = "task-1337569" [ 1059.430633] env[61006]: _type = "Task" [ 1059.430633] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.438155] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337569, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.571616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.571849] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.572106] env[61006]: DEBUG nova.objects.instance [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'resources' on Instance uuid 5c1c8651-cb82-49c7-9e0d-fe3960fc1633 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.746076] env[61006]: DEBUG oslo_concurrency.lockutils [req-2897d4cf-1bc0-413a-81c8-15167e333b2b req-58ae8010-99b7-4f11-bb61-81c607c70261 service nova] Releasing lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.940775] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337569, 'name': ReconfigVM_Task, 'duration_secs': 0.244511} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.941054] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b/volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.945670] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-271ddd93-216e-469b-b136-998eb04d0779 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.959661] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1059.959661] env[61006]: value = "task-1337570" [ 1059.959661] env[61006]: _type = "Task" [ 1059.959661] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.966898] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337570, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.152046] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1664a5d6-7577-4d58-837a-f5684ac80090 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.159472] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cd5ae7-9281-4516-9ffe-c50fcb5db41e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.190047] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bda574-1d37-400b-bf93-39ba0a6dd915 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.197369] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d92833-5eb9-4aa7-9f95-f20546012f55 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.210171] env[61006]: DEBUG nova.compute.provider_tree [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.469898] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337570, 'name': ReconfigVM_Task, 'duration_secs': 0.115924} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.470229] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285429', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'name': 'volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '059d8cea-e1d9-4e27-828f-e0256c83b1f1', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'serial': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1060.470764] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e31d5e8-a4ca-4ea8-b1a9-6bcd128ff244 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.477443] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1060.477443] env[61006]: value = "task-1337571" [ 1060.477443] env[61006]: _type = "Task" [ 1060.477443] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.485221] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337571, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.713216] env[61006]: DEBUG nova.scheduler.client.report [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.822396] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd7a845-903d-40d6-aa5f-d32bd035dc2c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.845777] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 0 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1060.989077] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337571, 'name': Rename_Task, 'duration_secs': 0.139035} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.991066] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.991066] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e54fdcd9-3d88-46e0-8cda-4600591e831a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.996432] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1060.996432] env[61006]: value = "task-1337572" [ 1060.996432] env[61006]: _type = "Task" [ 1060.996432] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.005212] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.218698] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.237286] env[61006]: INFO nova.scheduler.client.report [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted allocations for instance 5c1c8651-cb82-49c7-9e0d-fe3960fc1633 [ 1061.351664] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.351988] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ce61726-4dae-443d-8987-938adaeffa24 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.359969] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1061.359969] env[61006]: value = "task-1337573" [ 1061.359969] env[61006]: _type = "Task" [ 1061.359969] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.367982] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.508111] env[61006]: DEBUG oslo_vmware.api [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337572, 'name': PowerOnVM_Task, 'duration_secs': 0.476825} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.508111] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1061.508111] env[61006]: INFO nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Took 3.48 seconds to spawn the instance on the hypervisor. [ 1061.508111] env[61006]: DEBUG nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.509813] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b7d9dd-fa76-4825-913a-be1f57c9a147 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.746531] env[61006]: DEBUG oslo_concurrency.lockutils [None req-12e89a5e-bfcf-4b81-8482-32345d81a1c6 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "5c1c8651-cb82-49c7-9e0d-fe3960fc1633" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.216s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.870194] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337573, 'name': PowerOffVM_Task, 'duration_secs': 0.227396} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.870557] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.870667] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 17 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.028219] env[61006]: INFO nova.compute.manager [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Took 9.90 seconds to build instance. [ 1062.377567] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.377817] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.377969] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.378172] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.378324] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.378474] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.378692] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.378855] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.379035] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.379206] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.379380] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.384399] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d4ae664-78d1-4766-884e-892cadd53460 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.400825] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1062.400825] env[61006]: value = "task-1337574" [ 1062.400825] env[61006]: _type = "Task" [ 1062.400825] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.408668] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.469459] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.469770] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.529980] env[61006]: DEBUG oslo_concurrency.lockutils [None req-9d94cf93-1664-42bf-b6f3-1e0f066b3f30 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.413s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.910637] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337574, 'name': ReconfigVM_Task, 'duration_secs': 0.227801} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.910982] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 33 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.972765] env[61006]: INFO nova.compute.manager [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Detaching volume 5501507e-d513-49f9-9e05-434386d674ca [ 1063.007747] env[61006]: INFO nova.virt.block_device [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Attempting to driver detach volume 5501507e-d513-49f9-9e05-434386d674ca from mountpoint /dev/sdb [ 1063.008019] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1063.008226] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285417', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'name': 'volume-5501507e-d513-49f9-9e05-434386d674ca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '02408ca9-c580-444a-9608-a752146ca499', 'attached_at': '', 'detached_at': '', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'serial': '5501507e-d513-49f9-9e05-434386d674ca'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1063.009116] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1e2692-9716-4a07-b96a-520daabf7828 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.032469] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2864e630-36c8-4645-916b-d333d6d481a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.039275] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c7b39b-c82e-42ba-b207-5b3fc0edfc04 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.058713] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22021335-f5b1-4260-b4bc-8988ba80a733 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.073357] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] The volume has not been displaced from its original location: [datastore1] volume-5501507e-d513-49f9-9e05-434386d674ca/volume-5501507e-d513-49f9-9e05-434386d674ca.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1063.078454] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1063.078732] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17cb5ad8-442d-4cc9-93eb-8f22e6157719 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.096975] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1063.096975] env[61006]: value = "task-1337575" [ 1063.096975] env[61006]: _type = "Task" [ 1063.096975] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.104319] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.202446] env[61006]: DEBUG nova.compute.manager [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Received event network-changed-e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.202653] env[61006]: DEBUG nova.compute.manager [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Refreshing instance network info cache due to event network-changed-e41ece56-1df4-4013-bace-827b00f83fee. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1063.202870] env[61006]: DEBUG oslo_concurrency.lockutils [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] Acquiring lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.203155] env[61006]: DEBUG oslo_concurrency.lockutils [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] Acquired lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.203431] env[61006]: DEBUG nova.network.neutron [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Refreshing network info cache for port e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1063.417319] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.417539] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.417862] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.417952] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.418128] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.418314] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.418610] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.418842] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.418994] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.419181] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.419364] env[61006]: DEBUG nova.virt.hardware [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.424883] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1063.425194] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72c3deaa-d2b4-4542-9b58-3f28c5c19981 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.444283] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1063.444283] env[61006]: value = "task-1337576" [ 1063.444283] env[61006]: _type = "Task" [ 1063.444283] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.451978] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337576, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.609437] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337575, 'name': ReconfigVM_Task, 'duration_secs': 0.230694} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.609735] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1063.614340] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29b0b486-6f86-4fc1-831e-f97cc8e99c64 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.627909] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1063.627909] env[61006]: value = "task-1337577" [ 1063.627909] env[61006]: _type = "Task" [ 1063.627909] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.634923] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337577, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.909091] env[61006]: DEBUG nova.network.neutron [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updated VIF entry in instance network info cache for port e41ece56-1df4-4013-bace-827b00f83fee. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1063.909470] env[61006]: DEBUG nova.network.neutron [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating instance_info_cache with network_info: [{"id": "e41ece56-1df4-4013-bace-827b00f83fee", "address": "fa:16:3e:bb:c5:7b", "network": {"id": "352a1fb9-fc7f-44e2-a74b-8624e667091b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-539177150-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "440694a84a91482cb8363f0e38562833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41ece56-1d", "ovs_interfaceid": "e41ece56-1df4-4013-bace-827b00f83fee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.958777] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337576, 'name': ReconfigVM_Task, 'duration_secs': 0.172524} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.959266] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1063.960520] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2e4767-40be-439e-939e-e65d4e08cd06 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.003750] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.004226] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67717ef1-4558-42c2-aa36-b6cfbc66edf1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.034605] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1064.034605] env[61006]: value = "task-1337578" [ 1064.034605] env[61006]: _type = "Task" [ 1064.034605] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.042976] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337578, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.137987] env[61006]: DEBUG oslo_vmware.api [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337577, 'name': ReconfigVM_Task, 'duration_secs': 0.130144} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.138317] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285417', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'name': 'volume-5501507e-d513-49f9-9e05-434386d674ca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '02408ca9-c580-444a-9608-a752146ca499', 'attached_at': '', 'detached_at': '', 'volume_id': '5501507e-d513-49f9-9e05-434386d674ca', 'serial': '5501507e-d513-49f9-9e05-434386d674ca'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1064.411950] env[61006]: DEBUG oslo_concurrency.lockutils [req-07540c5c-f490-4d94-8958-b389b3b9beec req-5b80ebda-cf2c-4a29-8928-99386d818162 service nova] Releasing lock "refresh_cache-059d8cea-e1d9-4e27-828f-e0256c83b1f1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.546046] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337578, 'name': ReconfigVM_Task, 'duration_secs': 0.29213} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.546328] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.546612] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 50 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1064.679411] env[61006]: DEBUG nova.objects.instance [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid 02408ca9-c580-444a-9608-a752146ca499 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.053599] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdf78d1-ba97-4f69-b363-de741eaa3856 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.076427] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7036076-c27b-4c16-aa90-f56250e69fdf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.097256] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 67 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1065.687613] env[61006]: DEBUG oslo_concurrency.lockutils [None req-ce35e817-7a58-4f88-889b-be29838ddbde tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.217s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.714041] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.714041] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.714454] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "02408ca9-c580-444a-9608-a752146ca499-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.714454] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.714679] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.716783] env[61006]: INFO nova.compute.manager [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Terminating instance [ 1066.718643] env[61006]: DEBUG nova.compute.manager [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.718840] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.719694] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dbd597-86d4-4f6f-998b-df8d1b5484df {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.727366] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.727593] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73d1ebb9-60a5-4797-9d65-245512af0766 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.733566] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1066.733566] env[61006]: value = "task-1337579" [ 1066.733566] env[61006]: _type = "Task" [ 1066.733566] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.741672] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.742592] env[61006]: DEBUG nova.network.neutron [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Port cdb40136-9153-4611-9dc5-7bde778a2b5a binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1067.244132] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337579, 'name': PowerOffVM_Task, 'duration_secs': 0.191423} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.247747] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.247930] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.248368] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-527c382e-87dc-4f5a-a9eb-0cdda9402f19 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.307351] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.307504] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.307687] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleting the datastore file [datastore2] 02408ca9-c580-444a-9608-a752146ca499 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.307933] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65f0ee34-9cf8-405c-8699-c9a1ad68e0e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.314565] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1067.314565] env[61006]: value = "task-1337581" [ 1067.314565] env[61006]: _type = "Task" [ 1067.314565] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.322326] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337581, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.762761] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.763297] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.763297] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.825138] env[61006]: DEBUG oslo_vmware.api [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337581, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128168} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.825399] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.825600] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.825764] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.826043] env[61006]: INFO nova.compute.manager [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: 02408ca9-c580-444a-9608-a752146ca499] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1067.826263] env[61006]: DEBUG oslo.service.loopingcall [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.826461] env[61006]: DEBUG nova.compute.manager [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.826556] env[61006]: DEBUG nova.network.neutron [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.227744] env[61006]: DEBUG nova.compute.manager [req-370d11ba-0be4-4d71-8528-24c8e30c0208 req-991e3538-cc22-4124-80ff-72eab0c190c2 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Received event network-vif-deleted-aa62774d-28ae-41d6-8bbb-3dae4be79437 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.227907] env[61006]: INFO nova.compute.manager [req-370d11ba-0be4-4d71-8528-24c8e30c0208 req-991e3538-cc22-4124-80ff-72eab0c190c2 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Neutron deleted interface aa62774d-28ae-41d6-8bbb-3dae4be79437; detaching it from the instance and deleting it from the info cache [ 1068.228092] env[61006]: DEBUG nova.network.neutron [req-370d11ba-0be4-4d71-8528-24c8e30c0208 req-991e3538-cc22-4124-80ff-72eab0c190c2 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.707758] env[61006]: DEBUG nova.network.neutron [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.730841] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a71d6fe5-19b2-46c4-ac6f-279b906cd383 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.740716] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575c1b82-cff7-4005-8d3b-ce511598410a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.765127] env[61006]: DEBUG nova.compute.manager [req-370d11ba-0be4-4d71-8528-24c8e30c0208 req-991e3538-cc22-4124-80ff-72eab0c190c2 service nova] [instance: 02408ca9-c580-444a-9608-a752146ca499] Detach interface failed, port_id=aa62774d-28ae-41d6-8bbb-3dae4be79437, reason: Instance 02408ca9-c580-444a-9608-a752146ca499 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1068.795669] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.795901] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.796104] env[61006]: DEBUG nova.network.neutron [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.210687] env[61006]: INFO nova.compute.manager [-] [instance: 02408ca9-c580-444a-9608-a752146ca499] Took 1.38 seconds to deallocate network for instance. [ 1069.458925] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.459162] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.459306] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1069.509516] env[61006]: DEBUG nova.network.neutron [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.718060] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.718060] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.718060] env[61006]: DEBUG nova.objects.instance [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'resources' on Instance uuid 02408ca9-c580-444a-9608-a752146ca499 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.977314] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.977624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquired lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.977681] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 02408ca9-c580-444a-9608-a752146ca499] Forcefully refreshing network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1070.012251] env[61006]: DEBUG oslo_concurrency.lockutils [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.291216] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa964f3-2646-4eb6-8338-2b538b3af124 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.299094] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62871255-1221-480f-9226-41515b8c272e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.328767] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351f7ba8-7b36-4c9a-91ce-890e1e0f6fb5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.335711] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5150bc64-3e91-4b51-8fb7-44fd35ddebe7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.348382] env[61006]: DEBUG nova.compute.provider_tree [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.495015] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1070.521211] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155f4e37-c1cb-4bdd-a71c-c6db2a2ed652 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.528110] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4e8446-3477-463e-8623-4073e457507e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.851193] env[61006]: DEBUG nova.scheduler.client.report [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.067795] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.356138] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.638s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.375768] env[61006]: INFO nova.scheduler.client.report [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted allocations for instance 02408ca9-c580-444a-9608-a752146ca499 [ 1071.570108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Releasing lock "refresh_cache-02408ca9-c580-444a-9608-a752146ca499" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.570369] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 02408ca9-c580-444a-9608-a752146ca499] Updated the network info_cache for instance {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1071.570583] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.570752] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.570902] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.571112] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.571212] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.571369] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.571510] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1071.571698] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.616890] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d28d30-731d-4ab7-8910-4153b6031b81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.637986] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202b98ac-1c71-4906-925e-2a59055ea3a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.644831] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 83 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1071.884965] env[61006]: DEBUG oslo_concurrency.lockutils [None req-675d379c-ba6c-43fa-a9c8-6012e98c41d9 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "02408ca9-c580-444a-9608-a752146ca499" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.171s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.074618] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.074904] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.075034] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.075191] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1072.076129] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db432023-aaba-4307-8ea5-2ae7cae65543 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.083978] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578debe0-83ef-4f3f-99a5-f0d2071e44e5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.097813] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b23026-7e8f-4e10-b232-41e48176eb40 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.104164] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bad3d02-6bfb-437f-a339-65ef137ef37d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.133689] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180116MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1072.133842] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.134057] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.152053] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.152053] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d8fe555-ff7f-43d5-a7e0-85cf08af684d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.157670] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1072.157670] env[61006]: value = "task-1337582" [ 1072.157670] env[61006]: _type = "Task" [ 1072.157670] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.165263] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.667339] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337582, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.141227] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Applying migration context for instance c8d841a2-218a-48d7-8716-e47c29798b00 as it has an incoming, in-progress migration 7ee371a3-8cd6-48db-b4c1-032fe8217aac. Migration status is post-migrating {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1073.142020] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating resource usage from migration 7ee371a3-8cd6-48db-b4c1-032fe8217aac [ 1073.158044] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.158044] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.158044] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Migration 7ee371a3-8cd6-48db-b4c1-032fe8217aac is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1073.158044] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance c8d841a2-218a-48d7-8716-e47c29798b00 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.167988] env[61006]: DEBUG oslo_vmware.api [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337582, 'name': PowerOnVM_Task, 'duration_secs': 0.537253} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.168313] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.168551] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-200427e6-031e-4612-9ed6-a7b3297fac77 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance 'c8d841a2-218a-48d7-8716-e47c29798b00' progress to 100 {{(pid=61006) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1073.489591] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.489840] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.663709] env[61006]: INFO nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance bd11a05e-4147-41ef-b04b-34eb3a639be6 has allocations against this compute host but is not found in the database. [ 1073.663889] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1073.664051] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1073.732171] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04299823-5016-4974-9713-cde9acf25a8a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.739912] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b43de0-5d4e-4549-b31f-41ee2bbdd920 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.769618] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2fae7d-224e-401b-af17-480a6d05f966 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.776503] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84834d9b-5241-4c54-a6b2-fbf5e40762ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.788849] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.993044] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1074.292051] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.510028] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.796185] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1074.796387] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.662s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.796648] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.287s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.798147] env[61006]: INFO nova.compute.claims [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.800410] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.800555] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Cleaning up deleted instances {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1075.316781] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] There are 46 instances to clean {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1075.317110] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d7608c4b-a8ec-4e56-bcc3-165aff76f649] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.492616] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff5d191-d1bf-426c-955c-1d9b41df7e45 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.500445] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b3d938-3144-4654-806f-c518ec868fce {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.529150] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2941ebd2-bb88-410c-9989-25cf3560379d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.535858] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6efd7b-265b-459d-8b50-3161d903230b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.548342] env[61006]: DEBUG nova.compute.provider_tree [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.618535] env[61006]: DEBUG nova.network.neutron [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Port cdb40136-9153-4611-9dc5-7bde778a2b5a binding to destination host cpu-1 is already ACTIVE {{(pid=61006) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1075.618865] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.619227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.619482] env[61006]: DEBUG nova.network.neutron [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.822861] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 9486ebfc-4717-4c97-8b44-ac922737bc78] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.051928] env[61006]: DEBUG nova.scheduler.client.report [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.326448] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 5c1c8651-cb82-49c7-9e0d-fe3960fc1633] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.337183] env[61006]: DEBUG nova.network.neutron [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.556782] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.760s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.557346] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1076.834071] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a36c3e9b-83b9-434e-8fe1-601d389a8cf2] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.837792] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.062319] env[61006]: DEBUG nova.compute.utils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.063826] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1077.063989] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1077.100673] env[61006]: DEBUG nova.policy [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7c04e3804464cdc8c2925f7bbcf5369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4e87e7cb3f2444a8cc7bb7e65a99757', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1077.337724] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: feb3c68c-4c03-40cd-89a5-3310bfcab359] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.340465] env[61006]: DEBUG nova.compute.manager [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61006) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1077.553628] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Successfully created port: 42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1077.567384] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1077.843335] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 095f85df-7265-46a9-bdbb-963292384493] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.346641] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 3a3aa93a-818e-41d0-896e-72ba97b8b377] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.450254] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.450540] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.577081] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1078.602404] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.602679] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.602848] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.603044] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.603202] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.603356] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.603589] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.603765] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.603936] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.604122] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.604301] env[61006]: DEBUG nova.virt.hardware [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.605264] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3269fbae-264a-4074-99f5-a4063f0ae347 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.613986] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c86bd4-f207-458c-bb05-d9f6f87f327c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.849610] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0b2b6d52-8340-4f85-9fc0-ab1f2b079031] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.947383] env[61006]: DEBUG nova.compute.manager [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Received event network-vif-plugged-42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1078.947706] env[61006]: DEBUG oslo_concurrency.lockutils [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.947903] env[61006]: DEBUG oslo_concurrency.lockutils [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.948115] env[61006]: DEBUG oslo_concurrency.lockutils [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.948267] env[61006]: DEBUG nova.compute.manager [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] No waiting events found dispatching network-vif-plugged-42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1078.948439] env[61006]: WARNING nova.compute.manager [req-3bb3abcf-8c69-4a2a-9669-f90cdc5c7e1d req-953b3c97-efbc-49da-9cee-622cc56f5f6c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Received unexpected event network-vif-plugged-42fd2cd0-875e-4d4f-a739-79d68d2495e6 for instance with vm_state building and task_state spawning. [ 1078.952894] env[61006]: DEBUG nova.objects.instance [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'migration_context' on Instance uuid c8d841a2-218a-48d7-8716-e47c29798b00 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1079.034766] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Successfully updated port: 42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1079.353720] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d027fd07-c409-46e5-857f-cdd2c0479f53] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.529324] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fd24bd-cc55-4ba7-8883-7bea1742dc5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.537130] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f95f4c5-b1aa-45f0-80fb-0ce7bc613c5d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.540662] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.540821] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.540947] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.570920] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60af5ffa-5a25-4993-9470-da6f5da20874 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.583340] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d1bb84-0779-44f0-8781-13f30f71b657 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.597235] env[61006]: DEBUG nova.compute.provider_tree [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.857190] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: a1458c3b-f6d4-4cde-ad9a-4e4693ff312c] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.072993] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1080.077858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.077964] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.078105] env[61006]: INFO nova.compute.manager [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Shelving [ 1080.100199] env[61006]: DEBUG nova.scheduler.client.report [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.206151] env[61006]: DEBUG nova.network.neutron [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating instance_info_cache with network_info: [{"id": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "address": "fa:16:3e:07:4a:7b", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42fd2cd0-87", "ovs_interfaceid": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.360909] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 02408ca9-c580-444a-9608-a752146ca499] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.585113] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.585348] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d569eeb0-f0a3-4822-ad07-6ab09efd0d62 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.594035] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1080.594035] env[61006]: value = "task-1337583" [ 1080.594035] env[61006]: _type = "Task" [ 1080.594035] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.600795] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.708891] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.709328] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Instance network_info: |[{"id": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "address": "fa:16:3e:07:4a:7b", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42fd2cd0-87", "ovs_interfaceid": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1080.709861] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:4a:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42fd2cd0-875e-4d4f-a739-79d68d2495e6', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1080.718517] env[61006]: DEBUG oslo.service.loopingcall [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.718804] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1080.719597] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6176c955-937b-4f93-84a5-f291ddedbaae {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.740382] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1080.740382] env[61006]: value = "task-1337584" [ 1080.740382] env[61006]: _type = "Task" [ 1080.740382] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.748488] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337584, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.863999] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6de1f827-1a19-457b-8600-6546593e55ca] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.974480] env[61006]: DEBUG nova.compute.manager [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Received event network-changed-42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.974480] env[61006]: DEBUG nova.compute.manager [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Refreshing instance network info cache due to event network-changed-42fd2cd0-875e-4d4f-a739-79d68d2495e6. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1080.974568] env[61006]: DEBUG oslo_concurrency.lockutils [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] Acquiring lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.974711] env[61006]: DEBUG oslo_concurrency.lockutils [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] Acquired lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.974885] env[61006]: DEBUG nova.network.neutron [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Refreshing network info cache for port 42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1081.102749] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337583, 'name': PowerOffVM_Task, 'duration_secs': 0.267158} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.103047] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.103847] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd01518-12aa-42f7-8feb-b4d6ece4cd1b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.122142] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.672s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.128410] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb65768-8757-4e16-81bd-4479569cf7c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.250241] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337584, 'name': CreateVM_Task, 'duration_secs': 0.403097} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.250472] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1081.251108] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.251288] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.251608] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1081.251849] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8635606a-23d1-4132-b17b-3d408b0c847c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.256126] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1081.256126] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38180-6670-e327-7c1e-a1f71b542b9c" [ 1081.256126] env[61006]: _type = "Task" [ 1081.256126] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.263162] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38180-6670-e327-7c1e-a1f71b542b9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.367767] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: f048bd9c-048b-4439-bbbf-dfcee7f18d84] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.642933] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1081.643263] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-20a44608-17b9-4422-90fd-66348548aab0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.651243] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1081.651243] env[61006]: value = "task-1337585" [ 1081.651243] env[61006]: _type = "Task" [ 1081.651243] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.661234] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337585, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.664722] env[61006]: DEBUG nova.network.neutron [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updated VIF entry in instance network info cache for port 42fd2cd0-875e-4d4f-a739-79d68d2495e6. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1081.665085] env[61006]: DEBUG nova.network.neutron [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating instance_info_cache with network_info: [{"id": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "address": "fa:16:3e:07:4a:7b", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42fd2cd0-87", "ovs_interfaceid": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.767135] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52c38180-6670-e327-7c1e-a1f71b542b9c, 'name': SearchDatastore_Task, 'duration_secs': 0.009444} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.767538] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.767828] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1081.768667] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.768867] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.769080] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.769366] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29060525-fc3c-4903-a12c-42131bed1245 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.777744] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.777955] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.778709] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01c4c699-ff44-45dd-b6f1-cfc3179b6cd6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.785583] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1081.785583] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52681ea7-b5cf-7b7e-54ac-22c3ba513c4b" [ 1081.785583] env[61006]: _type = "Task" [ 1081.785583] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.792768] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52681ea7-b5cf-7b7e-54ac-22c3ba513c4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.870957] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 97d2dfbb-fba6-4b9b-82cd-8fda67d32ba1] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.161030] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337585, 'name': CreateSnapshot_Task, 'duration_secs': 0.429621} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.161307] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1082.162059] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb054b9-17e2-40da-bf48-239983cb3679 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.169091] env[61006]: DEBUG oslo_concurrency.lockutils [req-2801ee9e-f423-427f-96b7-c60ee4ff47a6 req-7a53ee24-11e4-4f2b-9d38-3b1f135df1e2 service nova] Releasing lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.295594] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52681ea7-b5cf-7b7e-54ac-22c3ba513c4b, 'name': SearchDatastore_Task, 'duration_secs': 0.008166} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.296382] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbd59250-e52a-4f7e-bca8-045a0da7f53f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.301307] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1082.301307] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ed8f6e-b937-55c5-915a-02b84e462653" [ 1082.301307] env[61006]: _type = "Task" [ 1082.301307] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.308438] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ed8f6e-b937-55c5-915a-02b84e462653, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.374478] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 40500a0d-33f3-492a-9573-ed10e45642c5] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.665638] env[61006]: INFO nova.compute.manager [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Swapping old allocation on dict_keys(['3360045e-46ab-4f2d-9377-dd481ab3cd53']) held by migration 7ee371a3-8cd6-48db-b4c1-032fe8217aac for instance [ 1082.680139] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1082.682611] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-46136b71-0793-4028-8264-22cf6650c76b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.691129] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1082.691129] env[61006]: value = "task-1337586" [ 1082.691129] env[61006]: _type = "Task" [ 1082.691129] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.694905] env[61006]: DEBUG nova.scheduler.client.report [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Overwriting current allocation {'allocations': {'3360045e-46ab-4f2d-9377-dd481ab3cd53': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 135}}, 'project_id': '08c673bf1b8a437fbfbfd34e912a8f37', 'user_id': '7707895279cb4c6ebc07bf3570f3f881', 'consumer_generation': 1} on consumer c8d841a2-218a-48d7-8716-e47c29798b00 {{(pid=61006) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1082.701033] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337586, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.785147] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.785379] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.785571] env[61006]: DEBUG nova.network.neutron [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.811512] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52ed8f6e-b937-55c5-915a-02b84e462653, 'name': SearchDatastore_Task, 'duration_secs': 0.008882} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.811822] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.812104] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd11a05e-4147-41ef-b04b-34eb3a639be6/bd11a05e-4147-41ef-b04b-34eb3a639be6.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.812376] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2fdebcb-0c90-4663-a28a-0270568b64ec {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.818561] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1082.818561] env[61006]: value = "task-1337587" [ 1082.818561] env[61006]: _type = "Task" [ 1082.818561] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.827312] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.878124] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: bd034bbc-80c5-4ed2-88d7-4a9396ba2c0a] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.203549] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337586, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.331390] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461917} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.331695] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd11a05e-4147-41ef-b04b-34eb3a639be6/bd11a05e-4147-41ef-b04b-34eb3a639be6.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.331935] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.332352] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c450eb6a-b199-4176-ac0e-a13ee276841b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.338610] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1083.338610] env[61006]: value = "task-1337588" [ 1083.338610] env[61006]: _type = "Task" [ 1083.338610] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.346354] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.381969] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 2259c42a-79f8-455f-b29f-464705526932] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.507319] env[61006]: DEBUG nova.network.neutron [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [{"id": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "address": "fa:16:3e:36:08:dd", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb40136-91", "ovs_interfaceid": "cdb40136-9153-4611-9dc5-7bde778a2b5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.702151] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337586, 'name': CloneVM_Task} progress is 95%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.847856] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069541} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.848094] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1083.848872] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d00eac-997a-449c-a0f0-90f7ab1965bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.870054] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] bd11a05e-4147-41ef-b04b-34eb3a639be6/bd11a05e-4147-41ef-b04b-34eb3a639be6.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.870299] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b311587-5935-4738-bef1-f2107548371e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.885610] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 1d6ede6a-5601-4da4-baa1-5090a112ec2d] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.889065] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1083.889065] env[61006]: value = "task-1337589" [ 1083.889065] env[61006]: _type = "Task" [ 1083.889065] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.896708] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.009931] env[61006]: DEBUG oslo_concurrency.lockutils [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-c8d841a2-218a-48d7-8716-e47c29798b00" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.010921] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b772796-b88d-4bbe-b1ae-38dedf9e1e79 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.018301] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335c31cf-5ce3-43cb-a8f6-b771e07eec1f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.203091] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337586, 'name': CloneVM_Task, 'duration_secs': 1.158903} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.203377] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Created linked-clone VM from snapshot [ 1084.204185] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a81636-5e06-485c-8708-9ccca3d90d43 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.211121] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Uploading image c168271f-0278-434f-8a71-1c70f9aadbfc {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1084.236492] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1084.236492] env[61006]: value = "vm-285436" [ 1084.236492] env[61006]: _type = "VirtualMachine" [ 1084.236492] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1084.236762] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ed71e733-9640-4180-bc86-bbfdb366cb1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.242757] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease: (returnval){ [ 1084.242757] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a88504-a5db-62dd-d066-aa8ac7f34831" [ 1084.242757] env[61006]: _type = "HttpNfcLease" [ 1084.242757] env[61006]: } obtained for exporting VM: (result){ [ 1084.242757] env[61006]: value = "vm-285436" [ 1084.242757] env[61006]: _type = "VirtualMachine" [ 1084.242757] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1084.243067] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the lease: (returnval){ [ 1084.243067] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a88504-a5db-62dd-d066-aa8ac7f34831" [ 1084.243067] env[61006]: _type = "HttpNfcLease" [ 1084.243067] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1084.248767] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1084.248767] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a88504-a5db-62dd-d066-aa8ac7f34831" [ 1084.248767] env[61006]: _type = "HttpNfcLease" [ 1084.248767] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1084.388641] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: b27fc628-1ac8-4283-bf6e-bcff1cbfe149] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.399306] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337589, 'name': ReconfigVM_Task, 'duration_secs': 0.350466} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.399587] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfigured VM instance instance-0000006c to attach disk [datastore2] bd11a05e-4147-41ef-b04b-34eb3a639be6/bd11a05e-4147-41ef-b04b-34eb3a639be6.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.400251] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48a95299-da9b-4e3b-8a0b-db07d4ebf490 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.406483] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1084.406483] env[61006]: value = "task-1337591" [ 1084.406483] env[61006]: _type = "Task" [ 1084.406483] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.414297] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337591, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.750824] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1084.750824] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a88504-a5db-62dd-d066-aa8ac7f34831" [ 1084.750824] env[61006]: _type = "HttpNfcLease" [ 1084.750824] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1084.751113] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1084.751113] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a88504-a5db-62dd-d066-aa8ac7f34831" [ 1084.751113] env[61006]: _type = "HttpNfcLease" [ 1084.751113] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1084.751823] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b057ba3f-2847-4096-aa59-39430e123be7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.758458] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1084.758693] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1084.844966] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-82244c71-16dd-47d7-b621-5dd9d8e1f121 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.894949] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 1a546f17-2fb8-4b99-9001-98cc6fe76837] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.915594] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337591, 'name': Rename_Task, 'duration_secs': 0.124109} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.915865] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.916639] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0799449d-eee4-4695-b2bb-796dd6756ff5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.922741] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1084.922741] env[61006]: value = "task-1337592" [ 1084.922741] env[61006]: _type = "Task" [ 1084.922741] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.930527] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337592, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.099518] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1085.100036] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3396a902-cb5c-4b8b-bd88-1e012bae3b01 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.106753] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1085.106753] env[61006]: value = "task-1337593" [ 1085.106753] env[61006]: _type = "Task" [ 1085.106753] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.115810] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337593, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.401189] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: c5d6f6df-d296-4d4d-815c-159dad2d08dc] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.433578] env[61006]: DEBUG oslo_vmware.api [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337592, 'name': PowerOnVM_Task, 'duration_secs': 0.448572} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.433824] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.434038] env[61006]: INFO nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Took 6.86 seconds to spawn the instance on the hypervisor. [ 1085.434226] env[61006]: DEBUG nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.434980] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41051eac-9d7c-441e-8623-b695fe3c8cdc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.617434] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337593, 'name': PowerOffVM_Task, 'duration_secs': 0.168094} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.617751] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.618494] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.618734] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.618912] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.619144] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.619326] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.619593] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.619850] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.619924] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.620139] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.620394] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.620545] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.626167] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-033903c1-7788-4f34-8ce2-168e699d57af {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.641588] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1085.641588] env[61006]: value = "task-1337594" [ 1085.641588] env[61006]: _type = "Task" [ 1085.641588] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.649966] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.904521] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: bdcf0445-6f8b-4e7b-bfb3-93a785162e56] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.950471] env[61006]: INFO nova.compute.manager [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Took 11.45 seconds to build instance. [ 1086.152819] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337594, 'name': ReconfigVM_Task, 'duration_secs': 0.339157} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.153306] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740a8493-aebd-4e43-a254-0413f5178093 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.175269] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1086.175540] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1086.175708] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.175899] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1086.176069] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.176227] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1086.176429] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1086.176592] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1086.176810] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1086.177023] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1086.177208] env[61006]: DEBUG nova.virt.hardware [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1086.178025] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbd9402e-a77a-4e32-9958-53fb6a4bab0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.183466] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1086.183466] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526efbaf-aa60-d568-5077-05c2fc1f5664" [ 1086.183466] env[61006]: _type = "Task" [ 1086.183466] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.191161] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526efbaf-aa60-d568-5077-05c2fc1f5664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.407706] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 84befce1-ccd4-4937-9656-e5cbb7882f47] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.452544] env[61006]: DEBUG oslo_concurrency.lockutils [None req-55927dac-f81f-4e70-ba65-3818b09b9d7b tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.962s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.564370] env[61006]: DEBUG nova.compute.manager [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Received event network-changed-42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.564570] env[61006]: DEBUG nova.compute.manager [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Refreshing instance network info cache due to event network-changed-42fd2cd0-875e-4d4f-a739-79d68d2495e6. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1086.564818] env[61006]: DEBUG oslo_concurrency.lockutils [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] Acquiring lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.565015] env[61006]: DEBUG oslo_concurrency.lockutils [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] Acquired lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.565148] env[61006]: DEBUG nova.network.neutron [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Refreshing network info cache for port 42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1086.694350] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]526efbaf-aa60-d568-5077-05c2fc1f5664, 'name': SearchDatastore_Task, 'duration_secs': 0.007071} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.700257] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1086.700985] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc47f43d-69b2-47d2-bbc2-7401890dab37 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.719861] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1086.719861] env[61006]: value = "task-1337595" [ 1086.719861] env[61006]: _type = "Task" [ 1086.719861] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.727838] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337595, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.911052] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d1e09ace-177d-430a-9e65-d01368815272] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.232656] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337595, 'name': ReconfigVM_Task, 'duration_secs': 0.405073} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.233149] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1087.233938] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90066028-6bb1-45ce-a3d9-8c0cbef7ee81 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.258735] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.259932] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6574ff8f-8312-4baf-a3a8-59a8e3427e0b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.279225] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1087.279225] env[61006]: value = "task-1337596" [ 1087.279225] env[61006]: _type = "Task" [ 1087.279225] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.288670] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337596, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.329509] env[61006]: DEBUG nova.network.neutron [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updated VIF entry in instance network info cache for port 42fd2cd0-875e-4d4f-a739-79d68d2495e6. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1087.331442] env[61006]: DEBUG nova.network.neutron [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating instance_info_cache with network_info: [{"id": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "address": "fa:16:3e:07:4a:7b", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42fd2cd0-87", "ovs_interfaceid": "42fd2cd0-875e-4d4f-a739-79d68d2495e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.414635] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 2232c77d-9f4e-4981-9f55-614bba5d71db] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.789763] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337596, 'name': ReconfigVM_Task, 'duration_secs': 0.305463} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.790513] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to attach disk [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00/c8d841a2-218a-48d7-8716-e47c29798b00.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.791488] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12d4bce-75d2-41eb-aca0-f13c10690e06 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.813578] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c100435f-9b6c-41a6-a1da-8e9fe983b296 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.849150] env[61006]: DEBUG oslo_concurrency.lockutils [req-b010feea-ebc6-468f-b846-a2454fdb4176 req-76d1d3b5-c5f2-4612-9b8d-5a06b9dfe50c service nova] Releasing lock "refresh_cache-bd11a05e-4147-41ef-b04b-34eb3a639be6" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.851230] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e69fa12-1818-4148-96b2-901e44bc7b34 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.881449] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf64d67-f4b9-450b-bcd7-15f15dd38a2b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.889025] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.890236] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f57ac4c9-72dd-4e7c-a9b6-2603ce67cb7b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.895348] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1087.895348] env[61006]: value = "task-1337597" [ 1087.895348] env[61006]: _type = "Task" [ 1087.895348] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.904124] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.918990] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 4d7bdc48-2ea5-449d-aaa8-a9986f0ece6e] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.406406] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337597, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.421244] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 66c93148-b91a-4d22-84af-f410c8e10875] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.907410] env[61006]: DEBUG oslo_vmware.api [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337597, 'name': PowerOnVM_Task, 'duration_secs': 0.578371} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.907697] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.924552] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: d174a8ec-867e-4fea-b878-2a9af1476949] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.427974] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 370f2153-adca-4513-8549-2bb7499cf913] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.930898] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 9c9fa347-bcfe-4009-af72-5f427e9d234a] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.954677] env[61006]: INFO nova.compute.manager [None req-84d7ddd8-33f4-4750-b140-a7d6c31287c9 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance to original state: 'active' [ 1090.434348] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42061ea3-d1d1-4633-bd24-65f7ee302c1f] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.937643] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 1a7b5a77-b43e-4c96-ac18-36634a0e7d5a] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.440546] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: aedaa5d6-e0f2-492c-a14b-3254863e1f06] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.530253] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.530621] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.530733] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.530932] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.531125] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.533319] env[61006]: INFO nova.compute.manager [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Terminating instance [ 1091.535365] env[61006]: DEBUG nova.compute.manager [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.535595] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.535852] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eef5a792-c9fe-4a68-a48b-81370fa4d9e6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.543437] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1091.543437] env[61006]: value = "task-1337598" [ 1091.543437] env[61006]: _type = "Task" [ 1091.543437] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.551825] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.943663] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 0111452e-1b4f-499c-932d-f31364d1a14c] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.053411] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337598, 'name': PowerOffVM_Task, 'duration_secs': 0.350582} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.053747] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.053991] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1092.054245] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285431', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'name': 'volume-37a60d5e-9349-4506-b970-81e4f379624f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c8d841a2-218a-48d7-8716-e47c29798b00', 'attached_at': '2024-10-22T16:18:06.000000', 'detached_at': '', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'serial': '37a60d5e-9349-4506-b970-81e4f379624f'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1092.055115] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dea632f-9a8f-4b5e-80b6-a1b26f7a9683 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.076505] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4983ee57-b403-4365-a703-237e39896d27 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.083757] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84e2bf5-1dcc-4161-8458-2950d1a5865b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.104015] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b489373f-34e5-4f7f-be95-a4a48f5260f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.118915] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] The volume has not been displaced from its original location: [datastore2] volume-37a60d5e-9349-4506-b970-81e4f379624f/volume-37a60d5e-9349-4506-b970-81e4f379624f.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1092.124342] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.124678] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1f8f45c-94b5-42f7-8371-465df76e9745 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.142978] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1092.142978] env[61006]: value = "task-1337599" [ 1092.142978] env[61006]: _type = "Task" [ 1092.142978] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.151127] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.447649] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 28b83ea7-5433-472d-9e47-f73a4f2fb389] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.653412] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337599, 'name': ReconfigVM_Task, 'duration_secs': 0.354075} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.653649] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.658278] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1e97448-a013-4f0a-9d0e-3f8ae30fb52e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.674068] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1092.674068] env[61006]: value = "task-1337600" [ 1092.674068] env[61006]: _type = "Task" [ 1092.674068] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.682469] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.951138] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 524f3fd1-1e71-40c0-96c2-0acac5055e01] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.184211] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337600, 'name': ReconfigVM_Task, 'duration_secs': 0.321299} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.184523] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285431', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'name': 'volume-37a60d5e-9349-4506-b970-81e4f379624f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c8d841a2-218a-48d7-8716-e47c29798b00', 'attached_at': '2024-10-22T16:18:06.000000', 'detached_at': '', 'volume_id': '37a60d5e-9349-4506-b970-81e4f379624f', 'serial': '37a60d5e-9349-4506-b970-81e4f379624f'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1093.184875] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.185649] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2458f83e-c0b2-4bdf-8957-62b189a925f1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.191875] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.192127] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d9f87eb-d362-4e17-9846-71bd3db4643a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.286197] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.286751] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.286876] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleting the datastore file [datastore2] c8d841a2-218a-48d7-8716-e47c29798b00 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.287207] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8079fc7-efd1-4c90-8150-69b9c88fe2e7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.294530] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1093.294530] env[61006]: value = "task-1337602" [ 1093.294530] env[61006]: _type = "Task" [ 1093.294530] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.302505] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.454314] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 72c3e37b-2eac-41d4-8308-0a6466c2dd24] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.805633] env[61006]: DEBUG oslo_vmware.api [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17045} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.805913] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.806091] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1093.806239] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1093.806412] env[61006]: INFO nova.compute.manager [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Took 2.27 seconds to destroy the instance on the hypervisor. [ 1093.806671] env[61006]: DEBUG oslo.service.loopingcall [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.806872] env[61006]: DEBUG nova.compute.manager [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1093.806969] env[61006]: DEBUG nova.network.neutron [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1093.957970] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 29c41817-2189-4622-8f35-86f61eb34bed] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.291512] env[61006]: DEBUG nova.compute.manager [req-827b2cb5-95b4-4686-bdd9-63f19c02ec56 req-13b9dd91-4f63-4224-b0d6-56e6e4d0aee3 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Received event network-vif-deleted-cdb40136-9153-4611-9dc5-7bde778a2b5a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.291512] env[61006]: INFO nova.compute.manager [req-827b2cb5-95b4-4686-bdd9-63f19c02ec56 req-13b9dd91-4f63-4224-b0d6-56e6e4d0aee3 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Neutron deleted interface cdb40136-9153-4611-9dc5-7bde778a2b5a; detaching it from the instance and deleting it from the info cache [ 1094.291992] env[61006]: DEBUG nova.network.neutron [req-827b2cb5-95b4-4686-bdd9-63f19c02ec56 req-13b9dd91-4f63-4224-b0d6-56e6e4d0aee3 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.461760] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 598c0d72-d679-49a8-b17c-f5f341c205e8] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.769354] env[61006]: DEBUG nova.network.neutron [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.794395] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8014684-f51c-45ae-b64b-a9dea5f79e2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.803961] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1649c665-b5b4-459f-9633-43870c372264 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.828974] env[61006]: DEBUG nova.compute.manager [req-827b2cb5-95b4-4686-bdd9-63f19c02ec56 req-13b9dd91-4f63-4224-b0d6-56e6e4d0aee3 service nova] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Detach interface failed, port_id=cdb40136-9153-4611-9dc5-7bde778a2b5a, reason: Instance c8d841a2-218a-48d7-8716-e47c29798b00 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1094.965111] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 42b92d52-d1f0-48ff-94b6-6164b387456c] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.275153] env[61006]: INFO nova.compute.manager [-] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Took 1.47 seconds to deallocate network for instance. [ 1095.471990] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 2a4089c9-4229-40bd-8d0e-706bba94655f] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.676544] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1095.677470] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18991b5-c8ef-4ba2-93a6-deb1cc1cb9e3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.683490] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1095.683657] env[61006]: ERROR oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk due to incomplete transfer. [ 1095.683945] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-151e2caa-315e-4a31-b8f0-a81828aa4687 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.690140] env[61006]: DEBUG oslo_vmware.rw_handles [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f26cd6-6f43-b8dc-4a3f-9698d445c8c6/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1095.690335] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Uploaded image c168271f-0278-434f-8a71-1c70f9aadbfc to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1095.692897] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1095.693139] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1116f6e0-b5f4-4a39-85d2-d52290ab44a0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.698357] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1095.698357] env[61006]: value = "task-1337603" [ 1095.698357] env[61006]: _type = "Task" [ 1095.698357] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.705690] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337603, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.822568] env[61006]: INFO nova.compute.manager [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: c8d841a2-218a-48d7-8716-e47c29798b00] Took 0.55 seconds to detach 1 volumes for instance. [ 1095.975647] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: ea7d5d49-ac76-4f2e-9456-912cf466fcc2] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.208469] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337603, 'name': Destroy_Task} progress is 33%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.329328] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.329643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.329818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.354674] env[61006]: INFO nova.scheduler.client.report [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted allocations for instance c8d841a2-218a-48d7-8716-e47c29798b00 [ 1096.479426] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 88e2bdc9-ab73-4e23-94b5-a45046835144] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.708599] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337603, 'name': Destroy_Task, 'duration_secs': 0.781428} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.708930] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Destroyed the VM [ 1096.709238] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1096.709521] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-37f9bb79-592c-4c16-84d9-e814daac505b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.716418] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1096.716418] env[61006]: value = "task-1337604" [ 1096.716418] env[61006]: _type = "Task" [ 1096.716418] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.724401] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337604, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.862085] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a2c2cfbb-fa0e-44e6-aa84-d3cf7fd7e4c8 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "c8d841a2-218a-48d7-8716-e47c29798b00" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.331s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.982393] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: e2a40bd9-fb66-40a2-bcf1-5c74707d59dd] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.226705] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337604, 'name': RemoveSnapshot_Task, 'duration_secs': 0.325977} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.226969] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1097.227268] env[61006]: DEBUG nova.compute.manager [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.228038] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363a7bfa-0fe4-44c8-b7e1-116415277fd2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.486291] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6e99894e-81b6-4a07-9ec7-caa16272b3ba] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.558432] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.558573] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.739369] env[61006]: INFO nova.compute.manager [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Shelve offloading [ 1097.741032] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.741316] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-270e0891-5553-4b2c-8979-26373b438738 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.748887] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1097.748887] env[61006]: value = "task-1337605" [ 1097.748887] env[61006]: _type = "Task" [ 1097.748887] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.756577] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337605, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.989721] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: cca6374c-09a4-4145-a116-c49c5a8330c1] Instance has had 0 of 5 cleanup attempts {{(pid=61006) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.060858] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1098.258547] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1098.258700] env[61006]: DEBUG nova.compute.manager [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.259526] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26088655-7c98-4866-960f-0e34473b7ff7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.265173] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.265338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.265507] env[61006]: DEBUG nova.network.neutron [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1098.492839] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.493229] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Cleaning up deleted instances with incomplete migration {{(pid=61006) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1098.586562] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.586823] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.588422] env[61006]: INFO nova.compute.claims [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.983859] env[61006]: DEBUG nova.network.neutron [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.994888] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1099.486234] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.661580] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3219c98-5c45-47ba-b40e-1f1448d8fb4f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.669550] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db5a1c0-8114-42fb-b6be-da53fbf60e04 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.699742] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c999ff-78c1-42a8-8bb2-ffdfdea554eb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.706636] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1838cd8-b817-46d6-9aad-f134f5bb37cd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.719151] env[61006]: DEBUG nova.compute.provider_tree [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.721214] env[61006]: INFO nova.compute.manager [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Rebuilding instance [ 1099.733554] env[61006]: DEBUG nova.compute.manager [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-vif-unplugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.733824] env[61006]: DEBUG oslo_concurrency.lockutils [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.734051] env[61006]: DEBUG oslo_concurrency.lockutils [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.734230] env[61006]: DEBUG oslo_concurrency.lockutils [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.734402] env[61006]: DEBUG nova.compute.manager [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] No waiting events found dispatching network-vif-unplugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1099.734566] env[61006]: WARNING nova.compute.manager [req-e145d7b9-3a44-436a-8d90-20f309436a97 req-f87fdbb2-4ab3-4c75-95e3-deaa5e37cbe9 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received unexpected event network-vif-unplugged-9b4cb734-5640-4e37-98ac-a329e9c9562d for instance with vm_state shelved and task_state shelving_offloading. [ 1099.763688] env[61006]: DEBUG nova.compute.manager [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.764064] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9ee0d1-4cac-45bd-81f2-5ecfeaf22ce4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.820033] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1099.820976] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2f7740-aab7-4463-8f95-b560749b2fa5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.828633] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.828866] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91036e30-3bae-42ff-953a-e7d6a621bc10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.886876] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.887116] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.887304] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.887555] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e75af178-e682-411d-b2ba-d2ef4844762c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.893845] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1099.893845] env[61006]: value = "task-1337607" [ 1099.893845] env[61006]: _type = "Task" [ 1099.893845] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.902697] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337607, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.223058] env[61006]: DEBUG nova.scheduler.client.report [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.274930] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.275227] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61ca5f92-5da4-47fe-8fb0-96ca1375ebf0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.282397] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1100.282397] env[61006]: value = "task-1337608" [ 1100.282397] env[61006]: _type = "Task" [ 1100.282397] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.290160] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.403828] env[61006]: DEBUG oslo_vmware.api [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337607, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140653} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.404050] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.404230] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.404448] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.424466] env[61006]: INFO nova.scheduler.client.report [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted allocations for instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 [ 1100.492180] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.728120] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.728683] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1100.791686] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337608, 'name': PowerOffVM_Task, 'duration_secs': 0.20211} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.791956] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1100.792632] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.792881] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f40d4b2-c43c-47fc-8d43-e0816ec33ca4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.798996] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1100.798996] env[61006]: value = "task-1337609" [ 1100.798996] env[61006]: _type = "Task" [ 1100.798996] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.806150] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.929365] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.929634] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.929858] env[61006]: DEBUG nova.objects.instance [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'resources' on Instance uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.234079] env[61006]: DEBUG nova.compute.utils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1101.235543] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1101.235710] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1101.272315] env[61006]: DEBUG nova.policy [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7707895279cb4c6ebc07bf3570f3f881', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c673bf1b8a437fbfbfd34e912a8f37', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1101.309816] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1101.310582] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1101.310582] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285429', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'name': 'volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '059d8cea-e1d9-4e27-828f-e0256c83b1f1', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'serial': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1101.311787] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47764d8-124b-4bae-944e-3b10dee9ed52 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.328749] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10c5540-c61a-46d2-aad6-14dad0563d16 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.335057] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60755350-8742-411e-bd2f-2c8fc292cf2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.352303] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bde4b95-f3e7-4075-8358-e3cf6e78ec71 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.366549] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] The volume has not been displaced from its original location: [datastore2] volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b/volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1101.371980] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Reconfiguring VM instance instance-0000006b to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.372282] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75d48288-e275-4eb7-87ac-9a4344e4b66b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.391379] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1101.391379] env[61006]: value = "task-1337610" [ 1101.391379] env[61006]: _type = "Task" [ 1101.391379] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.400818] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337610, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.432703] env[61006]: DEBUG nova.objects.instance [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'numa_topology' on Instance uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.534639] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Successfully created port: 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1101.738823] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1101.767150] env[61006]: DEBUG nova.compute.manager [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.767344] env[61006]: DEBUG nova.compute.manager [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing instance network info cache due to event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1101.767603] env[61006]: DEBUG oslo_concurrency.lockutils [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.768323] env[61006]: DEBUG oslo_concurrency.lockutils [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.768509] env[61006]: DEBUG nova.network.neutron [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1101.900972] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337610, 'name': ReconfigVM_Task, 'duration_secs': 0.184869} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.901277] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Reconfigured VM instance instance-0000006b to detach disk 2000 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.905832] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0cafb7c-3aaa-4577-8a69-4dc36dcf031b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.920970] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1101.920970] env[61006]: value = "task-1337611" [ 1101.920970] env[61006]: _type = "Task" [ 1101.920970] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.928873] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.935355] env[61006]: DEBUG nova.objects.base [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Object Instance<4d46a0be-247b-49e5-b8a4-54ae29b18218> lazy-loaded attributes: resources,numa_topology {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1101.987868] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec0ffbd-fe14-4b73-8923-edce28e6341c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.995372] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e3d87a-d889-4780-b053-c99145836576 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.024919] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426b49b8-0777-439e-b94e-e1a128fb0582 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.033856] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1939ebd7-0be0-4845-85de-58d560ed47b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.048119] env[61006]: DEBUG nova.compute.provider_tree [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.430965] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337611, 'name': ReconfigVM_Task, 'duration_secs': 0.107893} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.430965] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285429', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'name': 'volume-d3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '059d8cea-e1d9-4e27-828f-e0256c83b1f1', 'attached_at': '', 'detached_at': '', 'volume_id': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b', 'serial': 'd3b7bb17-e4e9-4255-a237-a79e5010ad0b'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1102.431276] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.431778] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe73bfb6-5141-467e-8162-cbe40baa9512 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.440384] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.440610] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-277b765f-59f5-4fcf-a37e-454e41f464a1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.510706] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1102.510896] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1102.511100] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Deleting the datastore file [datastore2] 059d8cea-e1d9-4e27-828f-e0256c83b1f1 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.511365] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8afa260-c626-4784-822d-689a03afb52e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.519488] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for the task: (returnval){ [ 1102.519488] env[61006]: value = "task-1337613" [ 1102.519488] env[61006]: _type = "Task" [ 1102.519488] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.524920] env[61006]: DEBUG nova.network.neutron [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updated VIF entry in instance network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1102.525316] env[61006]: DEBUG nova.network.neutron [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9b4cb734-56", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.530226] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337613, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.551178] env[61006]: DEBUG nova.scheduler.client.report [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.750886] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1102.775922] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1102.776202] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1102.776361] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1102.776541] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1102.776687] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1102.776835] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1102.777050] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1102.777215] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1102.777383] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1102.777542] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1102.777711] env[61006]: DEBUG nova.virt.hardware [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1102.778781] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9f191f-407b-4447-9fa6-3e1a6ede243b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.786396] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dafb82d-26b4-4b19-aa38-29dc4bf93f80 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.002177] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Successfully updated port: 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1103.028998] env[61006]: DEBUG oslo_vmware.api [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Task: {'id': task-1337613, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152794} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.029231] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.029412] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1103.029589] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1103.031449] env[61006]: DEBUG oslo_concurrency.lockutils [req-62187bc5-f8d1-41b3-8f42-7182f00ae3d8 req-24a1ce1e-6fa0-4609-b0e8-cb2f6f558f0c service nova] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.055938] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.126s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.079842] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1103.080160] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b1708da-84c3-48de-9597-72afcf71b23c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.089155] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fa357b-606b-4372-b343-2ee04f5d696a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.114771] env[61006]: ERROR nova.compute.manager [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Failed to detach volume d3b7bb17-e4e9-4255-a237-a79e5010ad0b from /dev/sda: nova.exception.InstanceNotFound: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Traceback (most recent call last): [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self.driver.rebuild(**kwargs) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise NotImplementedError() [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] NotImplementedError [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] During handling of the above exception, another exception occurred: [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Traceback (most recent call last): [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self.driver.detach_volume(context, old_connection_info, [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] return self._volumeops.detach_volume(connection_info, instance) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._detach_volume_vmdk(connection_info, instance) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] stable_ref.fetch_moref(session) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] nova.exception.InstanceNotFound: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. [ 1103.114771] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.235528] env[61006]: DEBUG nova.compute.utils [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Build of instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 aborted: Failed to rebuild volume backed instance. {{(pid=61006) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1103.237739] env[61006]: ERROR nova.compute.manager [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 aborted: Failed to rebuild volume backed instance. [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Traceback (most recent call last): [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self.driver.rebuild(**kwargs) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise NotImplementedError() [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] NotImplementedError [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] During handling of the above exception, another exception occurred: [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Traceback (most recent call last): [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._detach_root_volume(context, instance, root_bdm) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] with excutils.save_and_reraise_exception(): [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self.force_reraise() [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise self.value [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self.driver.detach_volume(context, old_connection_info, [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] return self._volumeops.detach_volume(connection_info, instance) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._detach_volume_vmdk(connection_info, instance) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] stable_ref.fetch_moref(session) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] nova.exception.InstanceNotFound: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] During handling of the above exception, another exception occurred: [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Traceback (most recent call last): [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] yield [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1103.237739] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._do_rebuild_instance_with_claim( [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._do_rebuild_instance( [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._rebuild_default_impl(**kwargs) [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] self._rebuild_volume_backed_instance( [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] raise exception.BuildAbortException( [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] nova.exception.BuildAbortException: Build of instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 aborted: Failed to rebuild volume backed instance. [ 1103.239367] env[61006]: ERROR nova.compute.manager [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] [ 1103.246589] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.504671] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.505027] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.505245] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1103.564163] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d56411d2-7621-4774-ad41-c4b432356fa1 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.486s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.565086] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.318s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.565250] env[61006]: INFO nova.compute.manager [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Unshelving [ 1103.795624] env[61006]: DEBUG nova.compute.manager [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.795923] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.796084] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.796349] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.796570] env[61006]: DEBUG nova.compute.manager [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] No waiting events found dispatching network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1103.796801] env[61006]: WARNING nova.compute.manager [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received unexpected event network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a for instance with vm_state building and task_state spawning. [ 1103.796975] env[61006]: DEBUG nova.compute.manager [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.797174] env[61006]: DEBUG nova.compute.manager [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing instance network info cache due to event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1103.797354] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.035804] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1104.171129] env[61006]: DEBUG nova.network.neutron [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.593704] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.594041] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.594259] env[61006]: DEBUG nova.objects.instance [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'pci_requests' on Instance uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.674147] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.674473] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance network_info: |[{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1104.675147] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.675354] env[61006]: DEBUG nova.network.neutron [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1104.676584] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:43:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b33352b-d3b8-47f4-a5a0-567fa5f8c16a', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1104.684058] env[61006]: DEBUG oslo.service.loopingcall [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.684923] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1104.685200] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3400f9aa-1eb8-452d-9863-14982be352a7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.704182] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1104.704182] env[61006]: value = "task-1337614" [ 1104.704182] env[61006]: _type = "Task" [ 1104.704182] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.711526] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337614, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.101038] env[61006]: DEBUG nova.objects.instance [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'numa_topology' on Instance uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.213565] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337614, 'name': CreateVM_Task, 'duration_secs': 0.275591} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.215671] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1105.216457] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.216536] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.216830] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1105.217376] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5134688-0ddd-4761-9a22-5f66e7af2938 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.221606] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1105.221606] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5281a95d-74f5-6708-9a81-11bee270188b" [ 1105.221606] env[61006]: _type = "Task" [ 1105.221606] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.228964] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5281a95d-74f5-6708-9a81-11bee270188b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.249522] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.369681] env[61006]: DEBUG nova.network.neutron [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updated VIF entry in instance network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1105.370125] env[61006]: DEBUG nova.network.neutron [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.571765] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.572058] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.572288] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.572478] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.572643] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.574952] env[61006]: INFO nova.compute.manager [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Terminating instance [ 1105.576833] env[61006]: DEBUG nova.compute.manager [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1105.577134] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e86a291e-cb1a-4332-9066-658e16fd3710 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.585623] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3347eed4-1709-40c6-8a33-f235ccd249c0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.602345] env[61006]: INFO nova.compute.claims [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.609797] env[61006]: WARNING nova.virt.vmwareapi.driver [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. [ 1105.609987] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.610427] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a0b05b8-2bf8-4c53-8e63-5fe481307b5f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.617710] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98aa486d-b821-4ce2-b00f-40fe8679e73f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.640907] env[61006]: WARNING nova.virt.vmwareapi.vmops [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. [ 1105.641273] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.641273] env[61006]: INFO nova.compute.manager [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1105.641512] env[61006]: DEBUG oslo.service.loopingcall [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.641728] env[61006]: DEBUG nova.compute.manager [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1105.641824] env[61006]: DEBUG nova.network.neutron [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1105.730986] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5281a95d-74f5-6708-9a81-11bee270188b, 'name': SearchDatastore_Task, 'duration_secs': 0.009697} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.731145] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.731379] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.731624] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.731782] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.731992] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.732268] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e4827d0-d398-4e8a-8b7c-9818c102f02c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.740326] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.740504] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1105.741189] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d37cd508-25dc-4419-9f2b-bb6861bbbfd9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.746379] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1105.746379] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0431b-53b2-21a1-6388-72fb2d461dcb" [ 1105.746379] env[61006]: _type = "Task" [ 1105.746379] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.755469] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0431b-53b2-21a1-6388-72fb2d461dcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.872892] env[61006]: DEBUG oslo_concurrency.lockutils [req-963cfc80-9c89-48c1-a42c-d550cd51e805 req-029787fb-4669-4496-a559-0fa25747ed25 service nova] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.097759] env[61006]: DEBUG nova.compute.manager [req-e833e104-388c-46cb-9f91-dd0a5adaef4e req-4ef286ce-77a4-4e22-9f71-6a0571b68abe service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Received event network-vif-deleted-e41ece56-1df4-4013-bace-827b00f83fee {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.097882] env[61006]: INFO nova.compute.manager [req-e833e104-388c-46cb-9f91-dd0a5adaef4e req-4ef286ce-77a4-4e22-9f71-6a0571b68abe service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Neutron deleted interface e41ece56-1df4-4013-bace-827b00f83fee; detaching it from the instance and deleting it from the info cache [ 1106.098063] env[61006]: DEBUG nova.network.neutron [req-e833e104-388c-46cb-9f91-dd0a5adaef4e req-4ef286ce-77a4-4e22-9f71-6a0571b68abe service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.257034] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52a0431b-53b2-21a1-6388-72fb2d461dcb, 'name': SearchDatastore_Task, 'duration_secs': 0.009907} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.257493] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74160c6d-084e-4833-8462-ed6da51f6726 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.262468] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1106.262468] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528cfd80-0e3c-53d8-3b7d-1ed7c6e42afb" [ 1106.262468] env[61006]: _type = "Task" [ 1106.262468] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.269968] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528cfd80-0e3c-53d8-3b7d-1ed7c6e42afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.568917] env[61006]: DEBUG nova.network.neutron [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.600339] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9f936ae-73e7-499d-ab88-6f483e748e0e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.609801] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cb818b-f509-49d4-aeea-79febefce38e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.638411] env[61006]: DEBUG nova.compute.manager [req-e833e104-388c-46cb-9f91-dd0a5adaef4e req-4ef286ce-77a4-4e22-9f71-6a0571b68abe service nova] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Detach interface failed, port_id=e41ece56-1df4-4013-bace-827b00f83fee, reason: Instance 059d8cea-e1d9-4e27-828f-e0256c83b1f1 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1106.691145] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7b600c-9e7f-4278-8dc8-fd26c45c1fbf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.699509] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b592b015-7f9b-472d-a4b7-7161f92149a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.735169] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cc7c05-20e8-49a7-8475-59df7e9f1b8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.743960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52bb344-280c-49bd-9373-62e4bb8c6474 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.758242] env[61006]: DEBUG nova.compute.provider_tree [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.772314] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]528cfd80-0e3c-53d8-3b7d-1ed7c6e42afb, 'name': SearchDatastore_Task, 'duration_secs': 0.010515} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.772628] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.773284] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1106.773284] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f9570b8-258e-4132-8b68-5c23aa9a3470 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.780338] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1106.780338] env[61006]: value = "task-1337615" [ 1106.780338] env[61006]: _type = "Task" [ 1106.780338] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.788283] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.071529] env[61006]: INFO nova.compute.manager [-] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Took 1.43 seconds to deallocate network for instance. [ 1107.261689] env[61006]: DEBUG nova.scheduler.client.report [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.290689] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46157} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.290987] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1107.291255] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1107.291474] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6fff069-7c52-44d0-95c7-a22055a314bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.297617] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1107.297617] env[61006]: value = "task-1337616" [ 1107.297617] env[61006]: _type = "Task" [ 1107.297617] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.305311] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.625200] env[61006]: INFO nova.compute.manager [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Took 0.55 seconds to detach 1 volumes for instance. [ 1107.627541] env[61006]: DEBUG nova.compute.manager [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Deleting volume: d3b7bb17-e4e9-4255-a237-a79e5010ad0b {{(pid=61006) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1107.766298] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.172s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.768461] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.519s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.809375] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062372} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.810094] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1107.810463] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4220fb-24c0-4f66-9309-13907e83efd1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.814121] env[61006]: INFO nova.network.neutron [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating port 9b4cb734-5640-4e37-98ac-a329e9c9562d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1107.834969] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.837977] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c2e0428-82fa-4b6a-bb10-68ca76e72f74 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.858556] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1107.858556] env[61006]: value = "task-1337618" [ 1107.858556] env[61006]: _type = "Task" [ 1107.858556] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.870736] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.884039] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d276178-1600-49cc-a8f1-544f2557de72 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.890971] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1046f96-f841-433e-8992-b19e2fce5a10 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.921432] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cf7345-8dc5-4c4b-a041-ed0e9feb4d8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.927716] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca95ef1-9947-44ac-9746-d978710e2db1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.940883] env[61006]: DEBUG nova.compute.provider_tree [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.168017] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.368713] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.444222] env[61006]: DEBUG nova.scheduler.client.report [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1108.870098] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337618, 'name': ReconfigVM_Task, 'duration_secs': 0.802078} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.870441] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.871099] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04cf1f84-5fd8-4408-835c-da3b6cf03f7e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.878088] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1108.878088] env[61006]: value = "task-1337619" [ 1108.878088] env[61006]: _type = "Task" [ 1108.878088] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.886246] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337619, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.949409] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.181s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.949672] env[61006]: INFO nova.compute.manager [None req-2e656678-1350-4f31-bee3-b68078087764 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] [instance: 059d8cea-e1d9-4e27-828f-e0256c83b1f1] Successfully reverted task state from rebuilding on failure for instance. [ 1108.955303] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.787s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.955535] env[61006]: DEBUG nova.objects.instance [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lazy-loading 'resources' on Instance uuid 059d8cea-e1d9-4e27-828f-e0256c83b1f1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.263127] env[61006]: DEBUG nova.compute.manager [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.263364] env[61006]: DEBUG oslo_concurrency.lockutils [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.263576] env[61006]: DEBUG oslo_concurrency.lockutils [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.263752] env[61006]: DEBUG oslo_concurrency.lockutils [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.263947] env[61006]: DEBUG nova.compute.manager [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] No waiting events found dispatching network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.264448] env[61006]: WARNING nova.compute.manager [req-1bd46e6b-3ee0-4e53-a240-4166c7bca9d3 req-d048d314-b8e9-4567-b139-546b1b279c8e service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received unexpected event network-vif-plugged-9b4cb734-5640-4e37-98ac-a329e9c9562d for instance with vm_state shelved_offloaded and task_state spawning. [ 1109.350059] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.350268] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.350474] env[61006]: DEBUG nova.network.neutron [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.387516] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337619, 'name': Rename_Task, 'duration_secs': 0.145402} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.387804] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1109.387921] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b044716a-b244-4ae3-95a4-df1d8180eb6b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.393825] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1109.393825] env[61006]: value = "task-1337620" [ 1109.393825] env[61006]: _type = "Task" [ 1109.393825] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.401047] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337620, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.507019] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.507256] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.529730] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17149e4f-8438-4063-9a7f-867f9d791199 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.536766] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774716da-7977-4565-8fb5-d95cd9c0faa1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.566800] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37dda086-ec53-47e3-910c-974e46aca2ba {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.573892] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd398b6-e23a-43e6-81b4-3cc96b0c2501 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.586549] env[61006]: DEBUG nova.compute.provider_tree [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.905971] env[61006]: DEBUG oslo_vmware.api [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337620, 'name': PowerOnVM_Task, 'duration_secs': 0.441807} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.906395] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1109.906842] env[61006]: INFO nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Took 7.16 seconds to spawn the instance on the hypervisor. [ 1109.907222] env[61006]: DEBUG nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1109.908467] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3058ee-7de3-4a34-89d8-ff7e2b13a4d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.009115] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1110.083198] env[61006]: DEBUG nova.network.neutron [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.089639] env[61006]: DEBUG nova.scheduler.client.report [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.426941] env[61006]: INFO nova.compute.manager [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Took 11.86 seconds to build instance. [ 1110.530633] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.586263] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.595155] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.640s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.597809] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.067s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.599241] env[61006]: INFO nova.compute.claims [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.621863] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f2625fdf044e2f919f9c4ec8334e17b5',container_format='bare',created_at=2024-10-22T16:18:01Z,direct_url=,disk_format='vmdk',id=c168271f-0278-434f-8a71-1c70f9aadbfc,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1048567806-shelved',owner='f60c05599bb7457f9bd7a3d11daf9ab3',properties=ImageMetaProps,protected=,size=31663104,status='active',tags=,updated_at=2024-10-22T16:18:17Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1110.621863] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1110.622069] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.622218] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1110.622367] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.622512] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1110.622716] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1110.622897] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1110.623085] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1110.623251] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1110.623500] env[61006]: DEBUG nova.virt.hardware [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1110.624542] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd156f0-ad9d-48c4-b1a2-8790c5fb8b98 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.633682] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebf76d4-134a-4333-ae53-c785dfe2a986 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.654453] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:87:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b4cb734-5640-4e37-98ac-a329e9c9562d', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.663158] env[61006]: DEBUG oslo.service.loopingcall [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.663998] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.664251] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04f2dc6e-c043-40f6-8d66-e20ce1413a7a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.683176] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.683176] env[61006]: value = "task-1337621" [ 1110.683176] env[61006]: _type = "Task" [ 1110.683176] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.690520] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337621, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.929222] env[61006]: DEBUG oslo_concurrency.lockutils [None req-7932e0db-362b-49c5-923c-b3f6e66a6086 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.370s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.116048] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a67ddb24-f905-4568-94c9-56171d8da1a0 tempest-ServerActionsV293TestJSON-169576442 tempest-ServerActionsV293TestJSON-169576442-project-member] Lock "059d8cea-e1d9-4e27-828f-e0256c83b1f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.544s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.193314] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337621, 'name': CreateVM_Task, 'duration_secs': 0.364402} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.193488] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1111.194238] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.194437] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.194834] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.195547] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b52e78c-68eb-4885-a51d-35f123de3bc6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.200907] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1111.200907] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525c0f88-bdda-8678-19d9-28f9e92afacf" [ 1111.200907] env[61006]: _type = "Task" [ 1111.200907] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.209950] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525c0f88-bdda-8678-19d9-28f9e92afacf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.288445] env[61006]: DEBUG nova.compute.manager [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.288638] env[61006]: DEBUG nova.compute.manager [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing instance network info cache due to event network-changed-9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1111.288848] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Acquiring lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.289014] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Acquired lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.289213] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Refreshing network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.670183] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4caa4f-0072-44a8-ac1d-55aa85ffef9f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.677719] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa18607c-412d-45ff-8b62-9be1d29afa87 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.709166] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c459021-9be3-4e12-97ff-81aec009ef8b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.718579] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4807c633-ae76-42c5-beb7-9d1775bf65a6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.722215] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.722435] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Processing image c168271f-0278-434f-8a71-1c70f9aadbfc {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.722675] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.722825] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.723015] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.723527] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1c0d324-5730-48df-910b-bb9d1d352ea1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.732361] env[61006]: DEBUG nova.compute.provider_tree [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.740392] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.740567] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.741768] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9546daf5-b456-42f3-97a6-bddf39fc91aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.747237] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1111.747237] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52528f3e-0a99-2fc4-6224-36bcd36ec5dd" [ 1111.747237] env[61006]: _type = "Task" [ 1111.747237] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.754620] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52528f3e-0a99-2fc4-6224-36bcd36ec5dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.984178] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updated VIF entry in instance network info cache for port 9b4cb734-5640-4e37-98ac-a329e9c9562d. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.984550] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [{"id": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "address": "fa:16:3e:53:87:75", "network": {"id": "284101b7-0673-4e43-967d-5902f6d17173", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1300732383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f60c05599bb7457f9bd7a3d11daf9ab3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4fc0575-c8e3-4f3c-b2e1-e10ac2d0cc1a", "external-id": "nsx-vlan-transportzone-256", "segmentation_id": 256, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b4cb734-56", "ovs_interfaceid": "9b4cb734-5640-4e37-98ac-a329e9c9562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.235882] env[61006]: DEBUG nova.scheduler.client.report [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.257895] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1112.258086] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Fetch image to [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0/OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1112.258270] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Downloading stream optimized image c168271f-0278-434f-8a71-1c70f9aadbfc to [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0/OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0.vmdk on the data store datastore2 as vApp {{(pid=61006) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1112.258444] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Downloading image file data c168271f-0278-434f-8a71-1c70f9aadbfc to the ESX as VM named 'OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0' {{(pid=61006) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1112.332191] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1112.332191] env[61006]: value = "resgroup-9" [ 1112.332191] env[61006]: _type = "ResourcePool" [ 1112.332191] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1112.332756] env[61006]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0cd1fb3d-437d-4a38-9529-e9d1e189373b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.352903] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease: (returnval){ [ 1112.352903] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bafdb9-d2a6-6eb3-f59d-7e239013bf87" [ 1112.352903] env[61006]: _type = "HttpNfcLease" [ 1112.352903] env[61006]: } obtained for vApp import into resource pool (val){ [ 1112.352903] env[61006]: value = "resgroup-9" [ 1112.352903] env[61006]: _type = "ResourcePool" [ 1112.352903] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1112.353225] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the lease: (returnval){ [ 1112.353225] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bafdb9-d2a6-6eb3-f59d-7e239013bf87" [ 1112.353225] env[61006]: _type = "HttpNfcLease" [ 1112.353225] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1112.359167] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.359167] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bafdb9-d2a6-6eb3-f59d-7e239013bf87" [ 1112.359167] env[61006]: _type = "HttpNfcLease" [ 1112.359167] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1112.487745] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Releasing lock "refresh_cache-4d46a0be-247b-49e5-b8a4-54ae29b18218" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.488018] env[61006]: DEBUG nova.compute.manager [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.488187] env[61006]: DEBUG nova.compute.manager [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing instance network info cache due to event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1112.488412] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.488551] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.488714] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.740538] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.143s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.741024] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1112.862076] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1112.862076] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bafdb9-d2a6-6eb3-f59d-7e239013bf87" [ 1112.862076] env[61006]: _type = "HttpNfcLease" [ 1112.862076] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1112.862076] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1112.862076] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52bafdb9-d2a6-6eb3-f59d-7e239013bf87" [ 1112.862076] env[61006]: _type = "HttpNfcLease" [ 1112.862076] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1112.862497] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4640aa4-0599-41a9-8cd4-be792fc7ca20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.869472] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1112.869718] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating HTTP connection to write to file with size = 31663104 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1112.932854] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c786dceb-7d40-4fb4-ba20-1bb1fe6ba65c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.210887] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updated VIF entry in instance network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1113.211346] env[61006]: DEBUG nova.network.neutron [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.246520] env[61006]: DEBUG nova.compute.utils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1113.250697] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Not allocating networking since 'none' was specified. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1113.716539] env[61006]: DEBUG oslo_concurrency.lockutils [req-d1e31927-bf95-4788-8729-28853a8b4bb0 req-1058836e-a3e3-4328-be86-89c12cafb195 service nova] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.752537] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1114.022398] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1114.022674] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1114.023721] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043f0004-a7df-4d05-8be4-76b36bf44b7a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.030819] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1114.031071] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1114.031374] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1316473f-5b9f-4408-8fac-c4b4399209d6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.223219] env[61006]: DEBUG oslo_vmware.rw_handles [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4986-ae48-270e-68f9-5d7b27c3a3a7/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1114.223555] env[61006]: INFO nova.virt.vmwareapi.images [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Downloaded image file data c168271f-0278-434f-8a71-1c70f9aadbfc [ 1114.224524] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af496a2c-bc19-477a-8277-023d1bac02c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.240077] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67e4d030-7766-4c29-beeb-a5f3739d3636 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.269957] env[61006]: INFO nova.virt.vmwareapi.images [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] The imported VM was unregistered [ 1114.271807] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1114.272049] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.272317] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf420dea-c339-418b-a026-b610746d05c6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.326883] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Created directory with path [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.327201] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0/OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0.vmdk to [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk. {{(pid=61006) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1114.327477] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-23f09fa4-8974-4ba1-9881-15884a09dab2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.333456] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1114.333456] env[61006]: value = "task-1337624" [ 1114.333456] env[61006]: _type = "Task" [ 1114.333456] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.340631] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.763123] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1114.844317] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.859148] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1114.859481] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1114.859651] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.859840] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1114.859993] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.860159] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1114.860370] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1114.860530] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1114.860698] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1114.860860] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1114.861046] env[61006]: DEBUG nova.virt.hardware [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.862016] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20791a9-c4fd-49af-ba65-3c6e9e096adc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.870070] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e48601-8e3e-4026-8c80-d4758d4cd479 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.883945] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Instance VIF info [] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1114.889860] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Creating folder: Project (6fed36ae226f40ffa2e01ce74a5c7d64). Parent ref: group-v285275. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1114.890244] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3036711-3f4e-480b-9547-487b17e619bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.900943] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Created folder: Project (6fed36ae226f40ffa2e01ce74a5c7d64) in parent group-v285275. [ 1114.901141] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Creating folder: Instances. Parent ref: group-v285440. {{(pid=61006) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1114.901400] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a15caada-67ea-4065-bb20-5989c0a1cd5d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.910954] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Created folder: Instances in parent group-v285440. [ 1114.911241] env[61006]: DEBUG oslo.service.loopingcall [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1114.911465] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1114.911703] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e1fef26-9c68-4dc5-8190-d2c17edba535 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.928474] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1114.928474] env[61006]: value = "task-1337627" [ 1114.928474] env[61006]: _type = "Task" [ 1114.928474] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.936208] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337627, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.345314] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.438964] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337627, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.845127] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.939925] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337627, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.347406] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.440972] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337627, 'name': CreateVM_Task} progress is 99%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.845772] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337624, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.30859} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.846060] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0/OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0.vmdk to [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk. [ 1116.846251] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Cleaning up location [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1116.846415] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_17aec5be-7dc5-4c5d-8736-911fdf5e78e0 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.846678] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b81091c4-337d-441f-94f5-b6bc39268a18 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.853040] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1116.853040] env[61006]: value = "task-1337628" [ 1116.853040] env[61006]: _type = "Task" [ 1116.853040] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.860212] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.941270] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337627, 'name': CreateVM_Task, 'duration_secs': 1.904843} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.941441] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.941897] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.942128] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.942444] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.942697] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7da53f46-8d1b-478e-84ed-7668962e2c02 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.946773] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1116.946773] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52112459-8b60-9741-6397-a6cb4be28c97" [ 1116.946773] env[61006]: _type = "Task" [ 1116.946773] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.955054] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52112459-8b60-9741-6397-a6cb4be28c97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.362554] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032642} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.362823] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.362989] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.363282] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk to [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1117.363532] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9058e2c6-ca7b-4828-b5fc-b9759fb66374 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.371084] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1117.371084] env[61006]: value = "task-1337629" [ 1117.371084] env[61006]: _type = "Task" [ 1117.371084] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.379278] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.456691] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52112459-8b60-9741-6397-a6cb4be28c97, 'name': SearchDatastore_Task, 'duration_secs': 0.008674} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.457015] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.457256] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.457491] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.457641] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.457812] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.458092] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5b8fccb-322b-46e6-8593-e59ba5d91505 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.465254] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.465439] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.466132] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa55444b-433b-4bd4-b12a-cfdf441b0f06 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.470739] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1117.470739] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5265096e-6cf8-eabe-e736-a9d792e0242c" [ 1117.470739] env[61006]: _type = "Task" [ 1117.470739] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.477870] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5265096e-6cf8-eabe-e736-a9d792e0242c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.882716] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.982205] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5265096e-6cf8-eabe-e736-a9d792e0242c, 'name': SearchDatastore_Task, 'duration_secs': 0.01014} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.983206] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b4f4205-a2b9-4b5d-8066-bcacedf6dbe2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.989046] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1117.989046] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5249d4fc-74c1-710d-950e-cc407ff68f10" [ 1117.989046] env[61006]: _type = "Task" [ 1117.989046] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.997234] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5249d4fc-74c1-710d-950e-cc407ff68f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.382959] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.499621] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5249d4fc-74c1-710d-950e-cc407ff68f10, 'name': SearchDatastore_Task, 'duration_secs': 0.080409} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.499893] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.500189] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a8270e7f-7b96-47b4-a973-6deca12a14a1/a8270e7f-7b96-47b4-a973-6deca12a14a1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1118.500477] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44b8946c-b65b-4b0c-b438-475127d0d0b3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.506774] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1118.506774] env[61006]: value = "task-1337630" [ 1118.506774] env[61006]: _type = "Task" [ 1118.506774] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.515346] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.885212] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.017357] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.383674] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.517523] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.884094] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337629, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.199097} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.884329] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c168271f-0278-434f-8a71-1c70f9aadbfc/c168271f-0278-434f-8a71-1c70f9aadbfc.vmdk to [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.885215] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80a58ec-8c51-4bfa-b457-208c92ad1649 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.908260] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.908780] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-629d36fd-a81b-4faf-82c9-980e1fdf5625 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.928419] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1119.928419] env[61006]: value = "task-1337631" [ 1119.928419] env[61006]: _type = "Task" [ 1119.928419] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.937166] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.018252] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337630, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.412895} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.018508] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] a8270e7f-7b96-47b4-a973-6deca12a14a1/a8270e7f-7b96-47b4-a973-6deca12a14a1.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1120.018717] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.018964] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-731295e9-8624-4540-b512-a03019ce4a89 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.025453] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1120.025453] env[61006]: value = "task-1337632" [ 1120.025453] env[61006]: _type = "Task" [ 1120.025453] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.032230] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.437974] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337631, 'name': ReconfigVM_Task, 'duration_secs': 0.291375} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.438280] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218/4d46a0be-247b-49e5-b8a4-54ae29b18218.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.438907] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-980ea786-c952-443d-9437-c0192d1d5e8f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.444611] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1120.444611] env[61006]: value = "task-1337633" [ 1120.444611] env[61006]: _type = "Task" [ 1120.444611] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.451596] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337633, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.534417] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062797} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.534680] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.535481] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bf1f63-e7fb-44a4-a1e6-c3b3c6d216ef {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.556220] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] a8270e7f-7b96-47b4-a973-6deca12a14a1/a8270e7f-7b96-47b4-a973-6deca12a14a1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.556536] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3009bf9e-1e68-4b57-ae27-831638dc8c96 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.575124] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1120.575124] env[61006]: value = "task-1337634" [ 1120.575124] env[61006]: _type = "Task" [ 1120.575124] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.582660] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.954833] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337633, 'name': Rename_Task, 'duration_secs': 0.1287} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.955312] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.955312] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8c7050a-7488-47ef-8af0-50861777844b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.961592] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1120.961592] env[61006]: value = "task-1337635" [ 1120.961592] env[61006]: _type = "Task" [ 1120.961592] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.969628] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.085645] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337634, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.471548] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337635, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.585916] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337634, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.972128] env[61006]: DEBUG oslo_vmware.api [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337635, 'name': PowerOnVM_Task, 'duration_secs': 0.833708} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.972510] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1122.068115] env[61006]: DEBUG nova.compute.manager [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.069091] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48148812-fd32-497e-a078-5ce55d51605c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.085720] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337634, 'name': ReconfigVM_Task, 'duration_secs': 1.132444} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.085978] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Reconfigured VM instance instance-0000006e to attach disk [datastore2] a8270e7f-7b96-47b4-a973-6deca12a14a1/a8270e7f-7b96-47b4-a973-6deca12a14a1.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.086560] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82d208ab-9ba0-43b4-80d5-4525c66c23e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.092356] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1122.092356] env[61006]: value = "task-1337636" [ 1122.092356] env[61006]: _type = "Task" [ 1122.092356] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.099694] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337636, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.589695] env[61006]: DEBUG oslo_concurrency.lockutils [None req-e06267ee-3704-4d98-b32f-faaef62f0f7e tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.024s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.601022] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337636, 'name': Rename_Task, 'duration_secs': 0.152496} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.601306] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.601546] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85cb0ddd-0917-4530-b3c2-7798e8f47dee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.607405] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1122.607405] env[61006]: value = "task-1337637" [ 1122.607405] env[61006]: _type = "Task" [ 1122.607405] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.615283] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337637, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.116548] env[61006]: DEBUG oslo_vmware.api [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337637, 'name': PowerOnVM_Task, 'duration_secs': 0.414352} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.116907] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.117017] env[61006]: INFO nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Took 8.35 seconds to spawn the instance on the hypervisor. [ 1123.117215] env[61006]: DEBUG nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1123.117952] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6b9f4b-e537-455a-9e46-b2b65b465d86 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.633429] env[61006]: INFO nova.compute.manager [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Took 13.12 seconds to build instance. [ 1124.135586] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d66f449-0ab9-4289-940b-60a5ac97d309 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.628s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.744796] env[61006]: DEBUG nova.compute.manager [None req-0e9614fe-198a-418f-9ac5-b1db353704a0 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.745776] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3ca113-c0c8-43ef-840d-a7220ffad42b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.806878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.807172] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.807361] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "a8270e7f-7b96-47b4-a973-6deca12a14a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.807550] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.807721] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.809757] env[61006]: INFO nova.compute.manager [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Terminating instance [ 1124.811322] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "refresh_cache-a8270e7f-7b96-47b4-a973-6deca12a14a1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.811483] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquired lock "refresh_cache-a8270e7f-7b96-47b4-a973-6deca12a14a1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.811649] env[61006]: DEBUG nova.network.neutron [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.966109] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.966359] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.255773] env[61006]: INFO nova.compute.manager [None req-0e9614fe-198a-418f-9ac5-b1db353704a0 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] instance snapshotting [ 1125.256441] env[61006]: DEBUG nova.objects.instance [None req-0e9614fe-198a-418f-9ac5-b1db353704a0 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lazy-loading 'flavor' on Instance uuid a8270e7f-7b96-47b4-a973-6deca12a14a1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.361454] env[61006]: DEBUG nova.network.neutron [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1125.414959] env[61006]: DEBUG nova.network.neutron [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.469591] env[61006]: DEBUG nova.compute.utils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1125.761960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bf134f-4890-4ebf-ac77-bf1c09d29164 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.788086] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37395d6d-f817-4fdf-b4ad-74a1be77b814 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.917458] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Releasing lock "refresh_cache-a8270e7f-7b96-47b4-a973-6deca12a14a1" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.917837] env[61006]: DEBUG nova.compute.manager [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1125.918055] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.918949] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b52acaa-536e-43ad-ae88-c8f54434715d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.926810] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.927053] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b32f967-ef82-4e02-9b1b-46b5424b8336 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.933402] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1125.933402] env[61006]: value = "task-1337638" [ 1125.933402] env[61006]: _type = "Task" [ 1125.933402] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.940677] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.972115] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.302515] env[61006]: DEBUG nova.compute.manager [None req-0e9614fe-198a-418f-9ac5-b1db353704a0 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Instance disappeared during snapshot {{(pid=61006) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 1126.416113] env[61006]: DEBUG nova.compute.manager [None req-0e9614fe-198a-418f-9ac5-b1db353704a0 tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Found 0 images (rotation: 2) {{(pid=61006) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1126.443768] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337638, 'name': PowerOffVM_Task, 'duration_secs': 0.117598} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.444060] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.444269] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.444523] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abaadae1-9c19-4232-9532-8fee56031533 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.468484] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.468762] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.468985] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Deleting the datastore file [datastore2] a8270e7f-7b96-47b4-a973-6deca12a14a1 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.469274] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e52926f-ec56-409d-9e05-cc3b4b5916bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.475896] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for the task: (returnval){ [ 1126.475896] env[61006]: value = "task-1337640" [ 1126.475896] env[61006]: _type = "Task" [ 1126.475896] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.483795] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.986115] env[61006]: DEBUG oslo_vmware.api [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Task: {'id': task-1337640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093267} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.986383] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.986573] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.986752] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.986929] env[61006]: INFO nova.compute.manager [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1126.987195] env[61006]: DEBUG oslo.service.loopingcall [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1126.987390] env[61006]: DEBUG nova.compute.manager [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1126.987485] env[61006]: DEBUG nova.network.neutron [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1127.002241] env[61006]: DEBUG nova.network.neutron [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1127.034910] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.035191] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.035423] env[61006]: INFO nova.compute.manager [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Attaching volume 272fdd07-6f39-4943-8966-b3b5d4d374b8 to /dev/sdb [ 1127.069567] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2082feff-3f88-43e2-986e-694a93c30f93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.076290] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2006d25-0b14-4d60-a03a-029f99253d77 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.088435] env[61006]: DEBUG nova.virt.block_device [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating existing volume attachment record: 9a4c76fd-a1be-453e-ab03-bdc64078dc5f {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1127.504992] env[61006]: DEBUG nova.network.neutron [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.008579] env[61006]: INFO nova.compute.manager [-] [instance: a8270e7f-7b96-47b4-a973-6deca12a14a1] Took 1.02 seconds to deallocate network for instance. [ 1128.459587] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.514949] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.515320] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.515489] env[61006]: DEBUG nova.objects.instance [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lazy-loading 'resources' on Instance uuid a8270e7f-7b96-47b4-a973-6deca12a14a1 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.077792] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705bb332-63c8-473e-94fd-78703591016f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.085430] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3430ead-9ed7-4d6f-bc08-1bfc0eb372bf {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.114817] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd49b96-2ceb-44a9-b11b-036b64952f66 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.122090] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e8604d-e1d3-442f-9d7d-0368d008cd7e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.136923] env[61006]: DEBUG nova.compute.provider_tree [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.459098] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.459388] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1129.640697] env[61006]: DEBUG nova.scheduler.client.report [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.146148] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.168401] env[61006]: INFO nova.scheduler.client.report [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Deleted allocations for instance a8270e7f-7b96-47b4-a973-6deca12a14a1 [ 1130.464608] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Didn't find any instances for network info cache update. {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1130.464864] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.676073] env[61006]: DEBUG oslo_concurrency.lockutils [None req-17da50b3-68d7-4a2c-8fe3-13164775b20d tempest-ServersAaction247Test-1397074325 tempest-ServersAaction247Test-1397074325-project-member] Lock "a8270e7f-7b96-47b4-a973-6deca12a14a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.869s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.968178] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.968382] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.968613] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.968801] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1130.969730] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2b8746-c0ea-4c70-944d-048617aca556 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.979642] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f62668-160b-4057-9fb8-f6d15d7d7076 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.995070] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6fa76d-7d70-47af-ae8d-ec8e44750ac2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.001682] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ed9f75-de8d-43b2-aad0-0ee468950228 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.031902] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180934MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1131.032068] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.032268] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.631520] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1131.631828] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285443', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'name': 'volume-272fdd07-6f39-4943-8966-b3b5d4d374b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd11a05e-4147-41ef-b04b-34eb3a639be6', 'attached_at': '', 'detached_at': '', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'serial': '272fdd07-6f39-4943-8966-b3b5d4d374b8'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1131.632850] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1363907-7aa3-49d0-ab65-2568ca154b7a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.650998] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845daf68-7f59-476f-a8c2-c303b49ee9de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.675134] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-272fdd07-6f39-4943-8966-b3b5d4d374b8/volume-272fdd07-6f39-4943-8966-b3b5d4d374b8.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.675722] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c7fb4dd-946d-43b5-ac6f-4066fd896b2e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.693148] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1131.693148] env[61006]: value = "task-1337643" [ 1131.693148] env[61006]: _type = "Task" [ 1131.693148] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.700526] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337643, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.058036] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance bd11a05e-4147-41ef-b04b-34eb3a639be6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1132.058206] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 6ecbaf80-a630-4ffc-949b-69759a2b64a3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1132.058327] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1132.058506] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1132.058643] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1132.105100] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a943cf77-dc6c-4bef-afff-8e029e8f2d9a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.112553] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b245d672-58ce-4640-b2e5-a194038af569 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.142472] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403d50f5-f744-41fd-869f-b83851c5c58c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.148952] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d928781-3b9d-48a2-8887-09215d011397 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.161440] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.202328] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337643, 'name': ReconfigVM_Task, 'duration_secs': 0.314519} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.202604] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-272fdd07-6f39-4943-8966-b3b5d4d374b8/volume-272fdd07-6f39-4943-8966-b3b5d4d374b8.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.207487] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed2e5b2-923c-4bfa-973c-5cb89c8f92b0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.221871] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1132.221871] env[61006]: value = "task-1337644" [ 1132.221871] env[61006]: _type = "Task" [ 1132.221871] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.229201] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.664146] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.731766] env[61006]: DEBUG oslo_vmware.api [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337644, 'name': ReconfigVM_Task, 'duration_secs': 0.124047} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.732082] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285443', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'name': 'volume-272fdd07-6f39-4943-8966-b3b5d4d374b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd11a05e-4147-41ef-b04b-34eb3a639be6', 'attached_at': '', 'detached_at': '', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'serial': '272fdd07-6f39-4943-8966-b3b5d4d374b8'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1133.169331] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1133.169528] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.137s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.766538] env[61006]: DEBUG nova.objects.instance [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid bd11a05e-4147-41ef-b04b-34eb3a639be6 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.163871] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.164067] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.164446] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.164610] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.164756] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.164895] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.271530] env[61006]: DEBUG oslo_concurrency.lockutils [None req-0d4e5294-f38c-46ca-927e-c4a8bac5bb28 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.236s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.476407] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.476684] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.668211] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Getting list of instances from cluster (obj){ [ 1134.668211] env[61006]: value = "domain-c8" [ 1134.668211] env[61006]: _type = "ClusterComputeResource" [ 1134.668211] env[61006]: } {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1134.670770] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8406493c-de1b-46cc-8bb2-e11f173dce63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.683202] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Got total of 3 instances {{(pid=61006) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1134.683360] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.683547] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid bd11a05e-4147-41ef-b04b-34eb3a639be6 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.683704] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Triggering sync for uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.684009] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.684227] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.684505] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.684726] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.684906] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.685115] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.685247] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1134.685887] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5a84cf-844e-40b0-8a5e-a9c7f34783a4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.688551] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc59971-5e0b-48e8-ba92-95b22370322d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.979552] env[61006]: INFO nova.compute.manager [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Detaching volume 272fdd07-6f39-4943-8966-b3b5d4d374b8 [ 1135.009279] env[61006]: INFO nova.virt.block_device [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Attempting to driver detach volume 272fdd07-6f39-4943-8966-b3b5d4d374b8 from mountpoint /dev/sdb [ 1135.009513] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1135.009705] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285443', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'name': 'volume-272fdd07-6f39-4943-8966-b3b5d4d374b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd11a05e-4147-41ef-b04b-34eb3a639be6', 'attached_at': '', 'detached_at': '', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'serial': '272fdd07-6f39-4943-8966-b3b5d4d374b8'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1135.010678] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608e9ad0-4eea-4c26-a652-fa942fd896f7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.031974] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d153a69-07c9-4466-a79d-5fd3f1d1cb56 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.038732] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e51ae3-4e84-4f8f-956f-9b3224f47a13 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.057812] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b991a17-ef3e-446b-af29-1233929eb545 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.071717] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] The volume has not been displaced from its original location: [datastore2] volume-272fdd07-6f39-4943-8966-b3b5d4d374b8/volume-272fdd07-6f39-4943-8966-b3b5d4d374b8.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1135.076753] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1135.077013] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24b85613-1767-41ce-965c-872241ebbbfd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.094462] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1135.094462] env[61006]: value = "task-1337645" [ 1135.094462] env[61006]: _type = "Task" [ 1135.094462] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.101652] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.198567] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.198975] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.604232] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337645, 'name': ReconfigVM_Task, 'duration_secs': 0.2092} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.604525] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1135.609103] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cee784b-b4f4-470c-bb15-97e9f7bb2990 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.623730] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1135.623730] env[61006]: value = "task-1337646" [ 1135.623730] env[61006]: _type = "Task" [ 1135.623730] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.633726] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.133514] env[61006]: DEBUG oslo_vmware.api [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337646, 'name': ReconfigVM_Task, 'duration_secs': 0.132471} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.133842] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285443', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'name': 'volume-272fdd07-6f39-4943-8966-b3b5d4d374b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd11a05e-4147-41ef-b04b-34eb3a639be6', 'attached_at': '', 'detached_at': '', 'volume_id': '272fdd07-6f39-4943-8966-b3b5d4d374b8', 'serial': '272fdd07-6f39-4943-8966-b3b5d4d374b8'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1136.675112] env[61006]: DEBUG nova.objects.instance [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid bd11a05e-4147-41ef-b04b-34eb3a639be6 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.682866] env[61006]: DEBUG oslo_concurrency.lockutils [None req-8591dd50-2793-4ef4-8101-b2473b1982ed tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.684023] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.999s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.687123] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458c081f-60e0-4d0e-91d3-b12942976ea1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.194845] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.705947] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.706292] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.706292] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.707096] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.707096] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.708742] env[61006]: INFO nova.compute.manager [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Terminating instance [ 1138.710544] env[61006]: DEBUG nova.compute.manager [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1138.710741] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1138.711584] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d34ca8-e629-479f-b4d5-58ec56eabb47 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.718960] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1138.719198] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ad2fe30-8740-413e-b9d4-3f8ae6c04a37 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.725678] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1138.725678] env[61006]: value = "task-1337647" [ 1138.725678] env[61006]: _type = "Task" [ 1138.725678] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.732613] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.235691] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337647, 'name': PowerOffVM_Task, 'duration_secs': 0.148456} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.235952] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1139.236140] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1139.236397] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a1edfd9-ff5e-4a61-8233-fd52663aca4e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.374352] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1139.374644] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1139.374876] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleting the datastore file [datastore2] bd11a05e-4147-41ef-b04b-34eb3a639be6 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1139.375189] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9cce7b1-af03-4414-8271-84f1a958dbfb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.381705] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1139.381705] env[61006]: value = "task-1337649" [ 1139.381705] env[61006]: _type = "Task" [ 1139.381705] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.389232] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.892534] env[61006]: DEBUG oslo_vmware.api [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159591} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.892902] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.892950] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1139.893149] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1139.893325] env[61006]: INFO nova.compute.manager [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1139.893569] env[61006]: DEBUG oslo.service.loopingcall [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.893769] env[61006]: DEBUG nova.compute.manager [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1139.893864] env[61006]: DEBUG nova.network.neutron [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1140.384782] env[61006]: DEBUG nova.compute.manager [req-435d2517-846d-48ad-b388-370cd2618a6e req-4fbfe15f-8f3c-4857-9a50-325fefb8062d service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Received event network-vif-deleted-42fd2cd0-875e-4d4f-a739-79d68d2495e6 {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.384988] env[61006]: INFO nova.compute.manager [req-435d2517-846d-48ad-b388-370cd2618a6e req-4fbfe15f-8f3c-4857-9a50-325fefb8062d service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Neutron deleted interface 42fd2cd0-875e-4d4f-a739-79d68d2495e6; detaching it from the instance and deleting it from the info cache [ 1140.385178] env[61006]: DEBUG nova.network.neutron [req-435d2517-846d-48ad-b388-370cd2618a6e req-4fbfe15f-8f3c-4857-9a50-325fefb8062d service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.842013] env[61006]: DEBUG nova.network.neutron [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.890893] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17eea045-2cf7-43c3-aacb-7571b49804be {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.899870] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cbcc43-4cf2-4287-976f-c0db01256a68 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.923858] env[61006]: DEBUG nova.compute.manager [req-435d2517-846d-48ad-b388-370cd2618a6e req-4fbfe15f-8f3c-4857-9a50-325fefb8062d service nova] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Detach interface failed, port_id=42fd2cd0-875e-4d4f-a739-79d68d2495e6, reason: Instance bd11a05e-4147-41ef-b04b-34eb3a639be6 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1141.344720] env[61006]: INFO nova.compute.manager [-] [instance: bd11a05e-4147-41ef-b04b-34eb3a639be6] Took 1.45 seconds to deallocate network for instance. [ 1141.851345] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.851617] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.851807] env[61006]: DEBUG nova.objects.instance [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'resources' on Instance uuid bd11a05e-4147-41ef-b04b-34eb3a639be6 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.404815] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f42134-cbef-4d75-92f5-159d8f269896 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.412136] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379f018c-4fe1-4eb8-aca8-66550bd6b6fb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.440709] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f92418c-7cbd-4372-91b7-c2a1cf3dd893 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.447573] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70ebd50-3959-43c9-a7a5-95ed8c00056d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.461532] env[61006]: DEBUG nova.compute.provider_tree [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.964597] env[61006]: DEBUG nova.scheduler.client.report [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.469897] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.488556] env[61006]: INFO nova.scheduler.client.report [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted allocations for instance bd11a05e-4147-41ef-b04b-34eb3a639be6 [ 1143.995970] env[61006]: DEBUG oslo_concurrency.lockutils [None req-c8c5f766-f31d-4163-b785-8130238614ad tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd11a05e-4147-41ef-b04b-34eb3a639be6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.290s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.511758] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.512066] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.014621] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Starting instance... {{(pid=61006) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1146.536587] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.536858] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.538348] env[61006]: INFO nova.compute.claims [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1147.593783] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31afa6e2-70ed-4b39-b063-155070c2f907 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.601220] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f726817-7606-488b-8ea1-8d3fbf73f82a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.630806] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131a23d5-7729-47ae-9166-9cb2797a67b2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.638156] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce13e40-6683-44ff-b33c-dd457ef4bded {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.650748] env[61006]: DEBUG nova.compute.provider_tree [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.996892] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.997199] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.997532] env[61006]: INFO nova.compute.manager [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Shelving [ 1148.153468] env[61006]: DEBUG nova.scheduler.client.report [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1148.505013] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1148.505272] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7c651b8-29ab-4501-a185-eedcd5b1e6cc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.512518] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1148.512518] env[61006]: value = "task-1337650" [ 1148.512518] env[61006]: _type = "Task" [ 1148.512518] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.520448] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.658818] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.122s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.659479] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Start building networks asynchronously for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1149.021902] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337650, 'name': PowerOffVM_Task, 'duration_secs': 0.230655} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.022185] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1149.022945] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7a90de-fb61-4e2e-95ca-6bdae26955e8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.040307] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f86831-0d5e-4849-b6f0-32cf0cd1caf9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.164184] env[61006]: DEBUG nova.compute.utils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.165608] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Allocating IP information in the background. {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1149.165783] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] allocate_for_instance() {{(pid=61006) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1149.208099] env[61006]: DEBUG nova.policy [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7c04e3804464cdc8c2925f7bbcf5369', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4e87e7cb3f2444a8cc7bb7e65a99757', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61006) authorize /opt/stack/nova/nova/policy.py:201}} [ 1149.453557] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Successfully created port: 7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1149.549984] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Creating Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1149.550358] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8fc815c9-d9af-491c-92cc-cc2f1d093fcd {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.559195] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1149.559195] env[61006]: value = "task-1337651" [ 1149.559195] env[61006]: _type = "Task" [ 1149.559195] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.567163] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337651, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.671498] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Start building block device mappings for instance. {{(pid=61006) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1150.069162] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337651, 'name': CreateSnapshot_Task, 'duration_secs': 0.46321} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.069438] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Created Snapshot of the VM instance {{(pid=61006) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1150.070310] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02600c61-80b0-4eb8-a6ea-9fa073ec91e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.587090] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Creating linked-clone VM from snapshot {{(pid=61006) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1150.587308] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d35ad383-ded1-4758-ad78-97f03fdec11b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.596815] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1150.596815] env[61006]: value = "task-1337652" [ 1150.596815] env[61006]: _type = "Task" [ 1150.596815] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.604383] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337652, 'name': CloneVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.677872] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Start spawning the instance on the hypervisor. {{(pid=61006) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1150.706715] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-22T16:06:39Z,direct_url=,disk_format='vmdk',id=2fb15142-70f7-4f09-b4ca-b3b9a23e35e0,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='6ca631e705b24eab98671274df919e64',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-22T16:06:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1150.707029] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1150.707203] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1150.707387] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1150.707536] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1150.707686] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1150.707891] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1150.708075] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1150.708271] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1150.708440] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1150.708619] env[61006]: DEBUG nova.virt.hardware [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1150.709515] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15706842-bcc9-4c27-8380-a3892eec60de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.716962] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5396fa72-9fe1-4d5c-abc9-cec6ab040532 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.834841] env[61006]: DEBUG nova.compute.manager [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Received event network-vif-plugged-7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1150.835134] env[61006]: DEBUG oslo_concurrency.lockutils [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.835392] env[61006]: DEBUG oslo_concurrency.lockutils [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.835532] env[61006]: DEBUG oslo_concurrency.lockutils [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.835738] env[61006]: DEBUG nova.compute.manager [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] No waiting events found dispatching network-vif-plugged-7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1150.835857] env[61006]: WARNING nova.compute.manager [req-6b99b6eb-2447-4e12-8e36-1b33ed7418dd req-8ebe5f78-2172-4483-913f-52fc20b4512c service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Received unexpected event network-vif-plugged-7e211d81-36db-43a5-b314-82c59336943a for instance with vm_state building and task_state spawning. [ 1150.968250] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Successfully updated port: 7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1151.106747] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337652, 'name': CloneVM_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.478666] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.478819] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.478971] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1151.607340] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337652, 'name': CloneVM_Task, 'duration_secs': 0.883804} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.607590] env[61006]: INFO nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Created linked-clone VM from snapshot [ 1151.608332] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b05616-1e40-4950-a014-d8fcc005f2c3 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.615500] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Uploading image 16bc28ff-faa0-44b0-a90b-b42477c71aeb {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1151.638204] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1151.638204] env[61006]: value = "vm-285445" [ 1151.638204] env[61006]: _type = "VirtualMachine" [ 1151.638204] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1151.638475] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7f4d2918-a37b-48ed-8e55-0fb049c0c499 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.645413] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease: (returnval){ [ 1151.645413] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b70be-5384-cae3-1747-44f1bb6cbb62" [ 1151.645413] env[61006]: _type = "HttpNfcLease" [ 1151.645413] env[61006]: } obtained for exporting VM: (result){ [ 1151.645413] env[61006]: value = "vm-285445" [ 1151.645413] env[61006]: _type = "VirtualMachine" [ 1151.645413] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1151.645672] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the lease: (returnval){ [ 1151.645672] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b70be-5384-cae3-1747-44f1bb6cbb62" [ 1151.645672] env[61006]: _type = "HttpNfcLease" [ 1151.645672] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1151.651560] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1151.651560] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b70be-5384-cae3-1747-44f1bb6cbb62" [ 1151.651560] env[61006]: _type = "HttpNfcLease" [ 1151.651560] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1152.011412] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1152.153474] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1152.153474] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b70be-5384-cae3-1747-44f1bb6cbb62" [ 1152.153474] env[61006]: _type = "HttpNfcLease" [ 1152.153474] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1152.153685] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1152.153685] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]525b70be-5384-cae3-1747-44f1bb6cbb62" [ 1152.153685] env[61006]: _type = "HttpNfcLease" [ 1152.153685] env[61006]: }. {{(pid=61006) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1152.154331] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d0c239-0e55-4453-8727-14386a4e7d9a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.161164] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1152.161339] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk for reading. {{(pid=61006) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1152.217656] env[61006]: DEBUG nova.network.neutron [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating instance_info_cache with network_info: [{"id": "7e211d81-36db-43a5-b314-82c59336943a", "address": "fa:16:3e:19:cb:5f", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e211d81-36", "ovs_interfaceid": "7e211d81-36db-43a5-b314-82c59336943a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.249324] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ff9e860d-478f-4f35-9117-481ffe4fde58 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.719967] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.720331] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Instance network_info: |[{"id": "7e211d81-36db-43a5-b314-82c59336943a", "address": "fa:16:3e:19:cb:5f", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e211d81-36", "ovs_interfaceid": "7e211d81-36db-43a5-b314-82c59336943a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61006) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1152.720769] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:cb:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e211d81-36db-43a5-b314-82c59336943a', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1152.729578] env[61006]: DEBUG oslo.service.loopingcall [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1152.729828] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1152.730075] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60956987-3e16-49cc-9674-91240ae05f1d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.751385] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1152.751385] env[61006]: value = "task-1337654" [ 1152.751385] env[61006]: _type = "Task" [ 1152.751385] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.862126] env[61006]: DEBUG nova.compute.manager [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Received event network-changed-7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1152.862424] env[61006]: DEBUG nova.compute.manager [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Refreshing instance network info cache due to event network-changed-7e211d81-36db-43a5-b314-82c59336943a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1152.862694] env[61006]: DEBUG oslo_concurrency.lockutils [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] Acquiring lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.862890] env[61006]: DEBUG oslo_concurrency.lockutils [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] Acquired lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.863120] env[61006]: DEBUG nova.network.neutron [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Refreshing network info cache for port 7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1153.260662] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337654, 'name': CreateVM_Task, 'duration_secs': 0.495209} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.262365] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1153.263184] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.263184] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.263344] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1153.263545] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e31d8dc-e649-4d00-9777-4cc989117882 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.268336] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1153.268336] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522453ff-019a-419f-3cfa-f24c7d6e360b" [ 1153.268336] env[61006]: _type = "Task" [ 1153.268336] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.276537] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522453ff-019a-419f-3cfa-f24c7d6e360b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.779304] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]522453ff-019a-419f-3cfa-f24c7d6e360b, 'name': SearchDatastore_Task, 'duration_secs': 0.009703} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.779761] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.780082] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Processing image 2fb15142-70f7-4f09-b4ca-b3b9a23e35e0 {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1153.780389] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.780593] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.780837] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1153.781174] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96b16872-1d09-4443-a441-8619daa90f45 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.789490] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1153.789688] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1153.790488] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99fd950d-459b-4786-8ca6-11d3ba1a53c9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.796463] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1153.796463] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5211f9d4-6e14-3305-aa81-e60680ca62b2" [ 1153.796463] env[61006]: _type = "Task" [ 1153.796463] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.804406] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5211f9d4-6e14-3305-aa81-e60680ca62b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.819764] env[61006]: DEBUG nova.network.neutron [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updated VIF entry in instance network info cache for port 7e211d81-36db-43a5-b314-82c59336943a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1153.820198] env[61006]: DEBUG nova.network.neutron [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating instance_info_cache with network_info: [{"id": "7e211d81-36db-43a5-b314-82c59336943a", "address": "fa:16:3e:19:cb:5f", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e211d81-36", "ovs_interfaceid": "7e211d81-36db-43a5-b314-82c59336943a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.306585] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5211f9d4-6e14-3305-aa81-e60680ca62b2, 'name': SearchDatastore_Task, 'duration_secs': 0.008639} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.307501] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aebac081-874f-44b2-b10f-26ab3a375d61 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.312907] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1154.312907] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52015923-05ab-6a79-d18b-df2c536d9d97" [ 1154.312907] env[61006]: _type = "Task" [ 1154.312907] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.321383] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52015923-05ab-6a79-d18b-df2c536d9d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.323363] env[61006]: DEBUG oslo_concurrency.lockutils [req-4a9fcb39-3c62-4997-b1bc-e034ea937463 req-57b0b0d6-e712-4831-8bd8-012395d5f1a9 service nova] Releasing lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.822954] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52015923-05ab-6a79-d18b-df2c536d9d97, 'name': SearchDatastore_Task, 'duration_secs': 0.009861} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.823236] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.823485] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd3aa1cb-f409-416b-9386-448b8bab5b17/bd3aa1cb-f409-416b-9386-448b8bab5b17.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1154.823784] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50c728c9-9150-4442-bfa6-0a9c0253c081 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.830742] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1154.830742] env[61006]: value = "task-1337655" [ 1154.830742] env[61006]: _type = "Task" [ 1154.830742] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.838440] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.341181] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337655, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.841783] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51177} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.842118] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0/2fb15142-70f7-4f09-b4ca-b3b9a23e35e0.vmdk to [datastore2] bd3aa1cb-f409-416b-9386-448b8bab5b17/bd3aa1cb-f409-416b-9386-448b8bab5b17.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1155.842393] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Extending root virtual disk to 1048576 {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1155.842706] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bdb30c06-ccbc-44d7-8713-8df9f61dbbd0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.849210] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1155.849210] env[61006]: value = "task-1337656" [ 1155.849210] env[61006]: _type = "Task" [ 1155.849210] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.856919] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.359376] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068226} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.359885] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Extended root virtual disk {{(pid=61006) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.360470] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e8cf95-9069-4bff-bd54-c5be754dcc7e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.382126] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] bd3aa1cb-f409-416b-9386-448b8bab5b17/bd3aa1cb-f409-416b-9386-448b8bab5b17.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.382464] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad8d59b7-bdfd-42f0-8e71-ae8f333d1b91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.403292] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1156.403292] env[61006]: value = "task-1337657" [ 1156.403292] env[61006]: _type = "Task" [ 1156.403292] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.410925] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337657, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.913459] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337657, 'name': ReconfigVM_Task, 'duration_secs': 0.300591} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.913755] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfigured VM instance instance-0000006f to attach disk [datastore2] bd3aa1cb-f409-416b-9386-448b8bab5b17/bd3aa1cb-f409-416b-9386-448b8bab5b17.vmdk or device None with type sparse {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.914441] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d23ac10-9881-403d-ace4-996f108389a5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.921220] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1156.921220] env[61006]: value = "task-1337658" [ 1156.921220] env[61006]: _type = "Task" [ 1156.921220] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.928929] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337658, 'name': Rename_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.431278] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337658, 'name': Rename_Task, 'duration_secs': 0.138895} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.431683] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.431799] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca57cf8d-a596-48b8-bacf-0be51c981539 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.438603] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1157.438603] env[61006]: value = "task-1337659" [ 1157.438603] env[61006]: _type = "Task" [ 1157.438603] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.446322] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.949381] env[61006]: DEBUG oslo_vmware.api [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337659, 'name': PowerOnVM_Task, 'duration_secs': 0.496704} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.949721] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1157.949956] env[61006]: INFO nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1157.950186] env[61006]: DEBUG nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1157.951050] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb7e6b2-a84e-4e32-baf1-171199bbefa6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.468110] env[61006]: INFO nova.compute.manager [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Took 11.95 seconds to build instance. [ 1158.823241] env[61006]: DEBUG nova.compute.manager [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Received event network-changed-7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.823374] env[61006]: DEBUG nova.compute.manager [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Refreshing instance network info cache due to event network-changed-7e211d81-36db-43a5-b314-82c59336943a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1158.823567] env[61006]: DEBUG oslo_concurrency.lockutils [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] Acquiring lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.823757] env[61006]: DEBUG oslo_concurrency.lockutils [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] Acquired lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.823946] env[61006]: DEBUG nova.network.neutron [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Refreshing network info cache for port 7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1158.971047] env[61006]: DEBUG oslo_concurrency.lockutils [None req-933afbc3-a47b-420d-b909-6e4aacc18085 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.458s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.542378] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.542792] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.542895] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.543104] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.543302] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.545700] env[61006]: DEBUG nova.network.neutron [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updated VIF entry in instance network info cache for port 7e211d81-36db-43a5-b314-82c59336943a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1159.546044] env[61006]: DEBUG nova.network.neutron [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating instance_info_cache with network_info: [{"id": "7e211d81-36db-43a5-b314-82c59336943a", "address": "fa:16:3e:19:cb:5f", "network": {"id": "71357338-97f6-435f-9798-10fa579b9027", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1751470860-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4e87e7cb3f2444a8cc7bb7e65a99757", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e211d81-36", "ovs_interfaceid": "7e211d81-36db-43a5-b314-82c59336943a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.547401] env[61006]: INFO nova.compute.manager [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Terminating instance [ 1159.549196] env[61006]: DEBUG nova.compute.manager [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1159.549394] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1159.550282] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c4d945-6c5f-480b-8e67-415b879fd498 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.559025] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1159.559295] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5636c338-9046-40c8-89fd-d4b7d72d582a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.566134] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1159.566134] env[61006]: value = "task-1337660" [ 1159.566134] env[61006]: _type = "Task" [ 1159.566134] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.574525] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.797545] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1159.798486] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb6a37b-d685-43c2-9bfe-2cb575ec623e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.806277] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1159.806444] env[61006]: ERROR oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk due to incomplete transfer. [ 1159.806661] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7244d854-5fda-4a22-8c89-e277f8c0d7d0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.812794] env[61006]: DEBUG oslo_vmware.rw_handles [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b1055-ebb9-0cdd-4f78-787d6dc56703/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1159.812995] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Uploaded image 16bc28ff-faa0-44b0-a90b-b42477c71aeb to the Glance image server {{(pid=61006) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1159.815375] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Destroying the VM {{(pid=61006) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1159.815615] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2b6b776e-d840-40c5-b273-7ef1829e9420 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.821125] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1159.821125] env[61006]: value = "task-1337661" [ 1159.821125] env[61006]: _type = "Task" [ 1159.821125] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.828502] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337661, 'name': Destroy_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.050670] env[61006]: DEBUG oslo_concurrency.lockutils [req-6e23b00d-bb53-4630-a762-0626a0a3069e req-e73e653c-fbd2-486e-9e55-9ee69ef361c2 service nova] Releasing lock "refresh_cache-bd3aa1cb-f409-416b-9386-448b8bab5b17" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.076307] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337660, 'name': PowerOffVM_Task, 'duration_secs': 0.231659} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.076602] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1160.076807] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1160.077082] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aac8c560-ba19-4ea5-840d-04fa6096d033 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.135114] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1160.135326] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1160.135508] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleting the datastore file [datastore2] 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1160.135783] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-700a1a8b-4950-441e-92e1-4b33e439cafc {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.142151] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for the task: (returnval){ [ 1160.142151] env[61006]: value = "task-1337663" [ 1160.142151] env[61006]: _type = "Task" [ 1160.142151] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.149204] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.330973] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337661, 'name': Destroy_Task, 'duration_secs': 0.306842} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.331269] env[61006]: INFO nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Destroyed the VM [ 1160.331510] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleting Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1160.331759] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-da82b997-ddde-4fcf-910e-6810cce8eb93 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.338340] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1160.338340] env[61006]: value = "task-1337664" [ 1160.338340] env[61006]: _type = "Task" [ 1160.338340] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.345698] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337664, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.651811] env[61006]: DEBUG oslo_vmware.api [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Task: {'id': task-1337663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456253} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.652205] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.652205] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1160.652350] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1160.652529] env[61006]: INFO nova.compute.manager [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1160.652859] env[61006]: DEBUG oslo.service.loopingcall [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1160.653095] env[61006]: DEBUG nova.compute.manager [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1160.653198] env[61006]: DEBUG nova.network.neutron [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1160.848822] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337664, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.172082] env[61006]: DEBUG nova.compute.manager [req-d08440f1-b1fd-40a0-b101-34b0c4b37708 req-ea72b60d-baa5-455f-b008-746f44446a91 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Received event network-vif-deleted-9b4cb734-5640-4e37-98ac-a329e9c9562d {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1161.172266] env[61006]: INFO nova.compute.manager [req-d08440f1-b1fd-40a0-b101-34b0c4b37708 req-ea72b60d-baa5-455f-b008-746f44446a91 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Neutron deleted interface 9b4cb734-5640-4e37-98ac-a329e9c9562d; detaching it from the instance and deleting it from the info cache [ 1161.172402] env[61006]: DEBUG nova.network.neutron [req-d08440f1-b1fd-40a0-b101-34b0c4b37708 req-ea72b60d-baa5-455f-b008-746f44446a91 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.349528] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337664, 'name': RemoveSnapshot_Task, 'duration_secs': 0.671063} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.349857] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleted Snapshot of the VM instance {{(pid=61006) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1161.350178] env[61006]: DEBUG nova.compute.manager [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1161.350936] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e17f0b1-df91-4311-a7ec-464767bb2d63 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.652644] env[61006]: DEBUG nova.network.neutron [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.674973] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ead9a99-9202-4a36-804b-e49d69cbb69c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.684220] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233f6b5c-0f3d-4c2f-815b-4ae84e4aee5d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.709339] env[61006]: DEBUG nova.compute.manager [req-d08440f1-b1fd-40a0-b101-34b0c4b37708 req-ea72b60d-baa5-455f-b008-746f44446a91 service nova] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Detach interface failed, port_id=9b4cb734-5640-4e37-98ac-a329e9c9562d, reason: Instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1161.863116] env[61006]: INFO nova.compute.manager [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Shelve offloading [ 1161.864757] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.865013] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df300a17-06b0-42cd-93e8-76b858c865bb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.872365] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1161.872365] env[61006]: value = "task-1337665" [ 1161.872365] env[61006]: _type = "Task" [ 1161.872365] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.879747] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.155455] env[61006]: INFO nova.compute.manager [-] [instance: 4d46a0be-247b-49e5-b8a4-54ae29b18218] Took 1.50 seconds to deallocate network for instance. [ 1162.382945] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] VM already powered off {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1162.383172] env[61006]: DEBUG nova.compute.manager [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1162.383960] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121933e4-4a1d-42d7-ae91-84a174030118 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.389560] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.389748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.389919] env[61006]: DEBUG nova.network.neutron [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.662475] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.662920] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.662920] env[61006]: DEBUG nova.objects.instance [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lazy-loading 'resources' on Instance uuid 4d46a0be-247b-49e5-b8a4-54ae29b18218 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.091225] env[61006]: DEBUG nova.network.neutron [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.214790] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4fc032-9f68-4737-a73f-7109607e23c5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.222578] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67c218-5e75-4e95-82ff-a93e9b56ec4d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.251832] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ec0316-233d-4c3a-83ea-e87963e208b4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.258778] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688a4f8c-3627-4dc8-8998-606ffc6d1af7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.272558] env[61006]: DEBUG nova.compute.provider_tree [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.594365] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.775481] env[61006]: DEBUG nova.scheduler.client.report [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.799288] env[61006]: DEBUG nova.compute.manager [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-vif-unplugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1163.799288] env[61006]: DEBUG oslo_concurrency.lockutils [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.799288] env[61006]: DEBUG oslo_concurrency.lockutils [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.799288] env[61006]: DEBUG oslo_concurrency.lockutils [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.799288] env[61006]: DEBUG nova.compute.manager [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] No waiting events found dispatching network-vif-unplugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1163.799288] env[61006]: WARNING nova.compute.manager [req-c4639d7b-944a-49f6-a3c5-30efb0965b14 req-756b6afb-8c16-4269-b4a9-dc0e1acb32f7 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received unexpected event network-vif-unplugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a for instance with vm_state shelved and task_state shelving_offloading. [ 1163.886285] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.887181] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e06a084-cec9-449a-9f6c-459b4c992c5a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.894496] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.894749] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ab865ce-4be0-4880-af9a-130f53b7bac5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.979280] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.979539] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.979807] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleting the datastore file [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.980164] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5662ac7-ef6c-4342-ac1f-41ca883d9408 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.987108] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1163.987108] env[61006]: value = "task-1337667" [ 1163.987108] env[61006]: _type = "Task" [ 1163.987108] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.994723] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.282122] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.300068] env[61006]: INFO nova.scheduler.client.report [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Deleted allocations for instance 4d46a0be-247b-49e5-b8a4-54ae29b18218 [ 1164.496226] env[61006]: DEBUG oslo_vmware.api [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14334} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.496520] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.496710] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.496891] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.513133] env[61006]: INFO nova.scheduler.client.report [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted allocations for instance 6ecbaf80-a630-4ffc-949b-69759a2b64a3 [ 1164.808215] env[61006]: DEBUG oslo_concurrency.lockutils [None req-2c72bb40-587b-42cf-af56-56c3309c3089 tempest-AttachVolumeShelveTestJSON-1649058296 tempest-AttachVolumeShelveTestJSON-1649058296-project-member] Lock "4d46a0be-247b-49e5-b8a4-54ae29b18218" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.265s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.017244] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.017538] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.017772] env[61006]: DEBUG nova.objects.instance [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'resources' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.521289] env[61006]: DEBUG nova.objects.instance [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'numa_topology' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.824469] env[61006]: DEBUG nova.compute.manager [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.825360] env[61006]: DEBUG nova.compute.manager [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing instance network info cache due to event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1165.825360] env[61006]: DEBUG oslo_concurrency.lockutils [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.825360] env[61006]: DEBUG oslo_concurrency.lockutils [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.825360] env[61006]: DEBUG nova.network.neutron [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1166.024158] env[61006]: DEBUG nova.objects.base [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Object Instance<6ecbaf80-a630-4ffc-949b-69759a2b64a3> lazy-loaded attributes: resources,numa_topology {{(pid=61006) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1166.060048] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14da4e00-7d58-45cc-8b89-eeea7738ac67 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.067420] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f902a554-122b-4c30-bd95-725baa88f507 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.096656] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9a1944-55ce-4952-a532-be1ccd999126 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.103807] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61ae200-750b-4861-8fc4-d97f6ed418fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.116365] env[61006]: DEBUG nova.compute.provider_tree [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1166.547721] env[61006]: DEBUG nova.network.neutron [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updated VIF entry in instance network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1166.547721] env[61006]: DEBUG nova.network.neutron [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": null, "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9b33352b-d3", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.619777] env[61006]: DEBUG nova.scheduler.client.report [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1166.892340] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.049222] env[61006]: DEBUG oslo_concurrency.lockutils [req-10c653f6-fb66-4e1f-8686-42b3fa72d95c req-bad19cf0-c34e-4375-a701-168a17c49329 service nova] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.124542] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.107s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.632283] env[61006]: DEBUG oslo_concurrency.lockutils [None req-07dfba75-495b-4a29-b6b1-0084ad8fce8a tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.635s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.633212] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.741s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.633322] env[61006]: INFO nova.compute.manager [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Unshelving [ 1168.661458] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.661802] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.662037] env[61006]: DEBUG nova.objects.instance [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'pci_requests' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.165869] env[61006]: DEBUG nova.objects.instance [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'numa_topology' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.668821] env[61006]: INFO nova.compute.claims [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1170.718259] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1cf177-384b-4c98-9717-eb001a161f57 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.725612] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a50517-3ae5-4dbb-86a1-a864de6d38f6 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.754616] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a110772b-1a81-4af0-bae2-ca799dbac67c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.761862] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d228abb4-5850-4eb1-9a8b-66e23ee1440f {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.776017] env[61006]: DEBUG nova.compute.provider_tree [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.278510] env[61006]: DEBUG nova.scheduler.client.report [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.783097] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.121s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.811714] env[61006]: INFO nova.network.neutron [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1173.202426] env[61006]: DEBUG nova.compute.manager [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1173.202676] env[61006]: DEBUG oslo_concurrency.lockutils [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.202881] env[61006]: DEBUG oslo_concurrency.lockutils [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.203202] env[61006]: DEBUG oslo_concurrency.lockutils [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.203405] env[61006]: DEBUG nova.compute.manager [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] No waiting events found dispatching network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1173.203577] env[61006]: WARNING nova.compute.manager [req-edee8edc-1cc4-4163-a13e-3d6701f2d40d req-0c1030d4-e619-4b31-911c-9b27ec0ee290 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received unexpected event network-vif-plugged-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a for instance with vm_state shelved_offloaded and task_state spawning. [ 1173.282616] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.282817] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.283223] env[61006]: DEBUG nova.network.neutron [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Building network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1173.973335] env[61006]: DEBUG nova.network.neutron [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1174.476391] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.502757] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-22T16:06:56Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='4a1f3338b831ef2a6947cd4ac1d0230c',container_format='bare',created_at=2024-10-22T16:19:09Z,direct_url=,disk_format='vmdk',id=16bc28ff-faa0-44b0-a90b-b42477c71aeb,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-408096984-shelved',owner='08c673bf1b8a437fbfbfd34e912a8f37',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-22T16:19:22Z,virtual_size=,visibility=), allow threads: False {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1174.503055] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1174.503257] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image limits 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1174.503477] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Flavor pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1174.503632] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Image pref 0:0:0 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1174.503784] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61006) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1174.503991] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1174.504170] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1174.504340] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Got 1 possible topologies {{(pid=61006) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1174.504538] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1174.504743] env[61006]: DEBUG nova.virt.hardware [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61006) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.505636] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e017144e-4a59-4c85-8b2f-a3f5efecea97 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.513669] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6002aaa5-e28b-4440-8ff6-5a65f9b04801 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.527243] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:43:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b33352b-d3b8-47f4-a5a0-567fa5f8c16a', 'vif_model': 'vmxnet3'}] {{(pid=61006) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1174.534977] env[61006]: DEBUG oslo.service.loopingcall [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1174.535240] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Creating VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1174.535452] env[61006]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46b348a7-91ea-4c60-8a3a-a42a88648d20 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.555561] env[61006]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1174.555561] env[61006]: value = "task-1337669" [ 1174.555561] env[61006]: _type = "Task" [ 1174.555561] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.563796] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337669, 'name': CreateVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.065922] env[61006]: DEBUG oslo_vmware.api [-] Task: {'id': task-1337669, 'name': CreateVM_Task, 'duration_secs': 0.309624} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.066119] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Created VM on the ESX host {{(pid=61006) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1175.066798] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.066971] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.067397] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1175.067649] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a22f6f5-fa5b-4e5e-8ecd-fb2c84ea488a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.072063] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1175.072063] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5221c41b-f6df-cf3f-8e5d-c1ddb5fd533e" [ 1175.072063] env[61006]: _type = "Task" [ 1175.072063] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.080899] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5221c41b-f6df-cf3f-8e5d-c1ddb5fd533e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.227282] env[61006]: DEBUG nova.compute.manager [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.227282] env[61006]: DEBUG nova.compute.manager [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing instance network info cache due to event network-changed-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1175.227282] env[61006]: DEBUG oslo_concurrency.lockutils [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.227466] env[61006]: DEBUG oslo_concurrency.lockutils [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.227584] env[61006]: DEBUG nova.network.neutron [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Refreshing network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1175.582958] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.583439] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Processing image 16bc28ff-faa0-44b0-a90b-b42477c71aeb {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1175.583531] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.583670] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.583849] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1175.584120] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91c87815-fe74-4ea5-a311-8093fd753cd1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.591761] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1175.591957] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61006) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1175.592640] env[61006]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac64a178-0755-4a9d-8c6d-e93a3db4c950 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.597132] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1175.597132] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52008820-ffb4-ba31-c674-cf0c70ee3308" [ 1175.597132] env[61006]: _type = "Task" [ 1175.597132] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.604297] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]52008820-ffb4-ba31-c674-cf0c70ee3308, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.923815] env[61006]: DEBUG nova.network.neutron [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updated VIF entry in instance network info cache for port 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a. {{(pid=61006) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1175.924208] env[61006]: DEBUG nova.network.neutron [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [{"id": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "address": "fa:16:3e:30:43:63", "network": {"id": "8dd7a44a-0665-4729-9496-fd1c8f8b9867", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-472001884-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c673bf1b8a437fbfbfd34e912a8f37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b33352b-d3", "ovs_interfaceid": "9b33352b-d3b8-47f4-a5a0-567fa5f8c16a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.106683] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Preparing fetch location {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1176.106904] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Fetch image to [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb/OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb.vmdk {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1176.107108] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Downloading stream optimized image 16bc28ff-faa0-44b0-a90b-b42477c71aeb to [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb/OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb.vmdk on the data store datastore2 as vApp {{(pid=61006) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1176.107285] env[61006]: DEBUG nova.virt.vmwareapi.images [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Downloading image file data 16bc28ff-faa0-44b0-a90b-b42477c71aeb to the ESX as VM named 'OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb' {{(pid=61006) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1176.171159] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1176.171159] env[61006]: value = "resgroup-9" [ 1176.171159] env[61006]: _type = "ResourcePool" [ 1176.171159] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1176.171437] env[61006]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1d8a8a01-7eef-402b-986a-c4c0d4b9d0f2 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.190528] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease: (returnval){ [ 1176.190528] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267b367-13cd-51fa-ce3c-f7ab699cf88d" [ 1176.190528] env[61006]: _type = "HttpNfcLease" [ 1176.190528] env[61006]: } obtained for vApp import into resource pool (val){ [ 1176.190528] env[61006]: value = "resgroup-9" [ 1176.190528] env[61006]: _type = "ResourcePool" [ 1176.190528] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1176.190826] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the lease: (returnval){ [ 1176.190826] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267b367-13cd-51fa-ce3c-f7ab699cf88d" [ 1176.190826] env[61006]: _type = "HttpNfcLease" [ 1176.190826] env[61006]: } to be ready. {{(pid=61006) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1176.197086] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.197086] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267b367-13cd-51fa-ce3c-f7ab699cf88d" [ 1176.197086] env[61006]: _type = "HttpNfcLease" [ 1176.197086] env[61006]: } is initializing. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1176.427718] env[61006]: DEBUG oslo_concurrency.lockutils [req-40334676-ca9d-4231-b07b-fc7f8eb0e9bb req-ba4793d3-bab9-4637-a538-e91e2c4209c8 service nova] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.699074] env[61006]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.699074] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267b367-13cd-51fa-ce3c-f7ab699cf88d" [ 1176.699074] env[61006]: _type = "HttpNfcLease" [ 1176.699074] env[61006]: } is ready. {{(pid=61006) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1176.699553] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1176.699553] env[61006]: value = "session[524e0b49-f405-ff2c-9a77-42e0e2c3b7d4]5267b367-13cd-51fa-ce3c-f7ab699cf88d" [ 1176.699553] env[61006]: _type = "HttpNfcLease" [ 1176.699553] env[61006]: }. {{(pid=61006) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1176.700511] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283d12ac-8fad-4808-a11c-333d02440e1c {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.707386] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk from lease info. {{(pid=61006) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1176.707562] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk. {{(pid=61006) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1176.769751] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e52e4df5-9737-4fb9-ab6b-b7bb6367e848 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.860068] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Completed reading data from the image iterator. {{(pid=61006) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1177.860464] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1177.861285] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd3b3a5-ae8d-4ea7-84b6-8c434b57c995 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.867796] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk is in state: ready. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1177.867985] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk. {{(pid=61006) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1177.868291] env[61006]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-55a63f4c-f362-424e-9ba3-956d41f52000 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.028786] env[61006]: DEBUG oslo_vmware.rw_handles [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fe2563-477a-1c4f-5a8e-b38202949d23/disk-0.vmdk. {{(pid=61006) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1178.029045] env[61006]: INFO nova.virt.vmwareapi.images [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Downloaded image file data 16bc28ff-faa0-44b0-a90b-b42477c71aeb [ 1178.030124] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d670881a-b945-4df1-bcf4-31a8ffe30c95 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.045270] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17f11520-6d13-4047-9f92-d481c8dd39c1 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.073179] env[61006]: INFO nova.virt.vmwareapi.images [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] The imported VM was unregistered [ 1178.075608] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Caching image {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1178.075846] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Creating directory with path [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.076122] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62a9cfdf-124a-4baf-8f4f-cc6f0bdcdeb0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.093351] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Created directory with path [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb {{(pid=61006) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.093471] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb/OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb.vmdk to [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk. {{(pid=61006) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1178.093788] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d6bba4f9-7208-4f54-b5f6-c80012f818ea {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.100317] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1178.100317] env[61006]: value = "task-1337672" [ 1178.100317] env[61006]: _type = "Task" [ 1178.100317] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.109656] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.610631] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.111797] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.612452] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.113318] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.615089] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337672, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.265458} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.615089] env[61006]: INFO nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb/OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb.vmdk to [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk. [ 1180.615089] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Cleaning up location [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb {{(pid=61006) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1180.615089] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9e555dd8-243c-434a-8a55-c4e391d4c2bb {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1180.615426] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-291c181b-ee7a-4b7e-bbbb-e9045a677844 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.621166] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1180.621166] env[61006]: value = "task-1337673" [ 1180.621166] env[61006]: _type = "Task" [ 1180.621166] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.628727] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.131404] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294769} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.131783] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1181.131873] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.132169] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk to [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1181.132442] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09bbff04-2503-480e-873c-4a0cad89b12a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.139884] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1181.139884] env[61006]: value = "task-1337674" [ 1181.139884] env[61006]: _type = "Task" [ 1181.139884] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.147423] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.649690] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.150441] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.651337] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.152049] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.653274] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.153251] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337674, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.775781} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.153645] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/16bc28ff-faa0-44b0-a90b-b42477c71aeb/16bc28ff-faa0-44b0-a90b-b42477c71aeb.vmdk to [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk {{(pid=61006) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1184.154324] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c12d165-f928-4748-bb3d-50aabb4b7dee {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.175172] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.175567] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-929a87c3-116e-4ce2-8522-7e04d208a1d5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.194298] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1184.194298] env[61006]: value = "task-1337675" [ 1184.194298] env[61006]: _type = "Task" [ 1184.194298] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.201645] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337675, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.705220] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337675, 'name': ReconfigVM_Task, 'duration_secs': 0.287593} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.705470] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3/6ecbaf80-a630-4ffc-949b-69759a2b64a3.vmdk or device None with type streamOptimized {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1184.706071] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62651941-8ec9-43a4-bb69-cc9d76a3a2b8 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.711606] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1184.711606] env[61006]: value = "task-1337676" [ 1184.711606] env[61006]: _type = "Task" [ 1184.711606] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.718734] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337676, 'name': Rename_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.222053] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337676, 'name': Rename_Task, 'duration_secs': 0.12902} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.222355] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powering on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1185.222607] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-603268d7-0158-40c3-bbf2-7dc03cb43a92 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.228834] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1185.228834] env[61006]: value = "task-1337677" [ 1185.228834] env[61006]: _type = "Task" [ 1185.228834] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.236239] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337677, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.738606] env[61006]: DEBUG oslo_vmware.api [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337677, 'name': PowerOnVM_Task, 'duration_secs': 0.40951} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.738975] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powered on the VM {{(pid=61006) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1185.837965] env[61006]: DEBUG nova.compute.manager [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Checking state {{(pid=61006) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.838944] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894d97c4-8b72-484f-933f-53a893598367 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.356888] env[61006]: DEBUG oslo_concurrency.lockutils [None req-83a1070c-b1f3-4d42-96cb-58b10c733131 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.724s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.646561] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.646813] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.647018] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.647221] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.647398] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.649559] env[61006]: INFO nova.compute.manager [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Terminating instance [ 1186.651370] env[61006]: DEBUG nova.compute.manager [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1186.651567] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.652394] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3108e8a8-8177-4a5d-a1ed-0433b7a751e0 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.660244] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1186.660724] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2dc954f-b2c5-48cd-8094-2e215ec97465 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.667157] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1186.667157] env[61006]: value = "task-1337678" [ 1186.667157] env[61006]: _type = "Task" [ 1186.667157] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.675369] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.177154] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337678, 'name': PowerOffVM_Task, 'duration_secs': 0.16279} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.177422] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1187.177592] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.177832] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10a2cf11-aecc-4af0-9287-543d0ca92c3e {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.233178] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.233414] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.233613] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleting the datastore file [datastore2] 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.233877] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e08f9d15-5212-45c7-9a96-ccab7fe8b4db {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.239844] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for the task: (returnval){ [ 1187.239844] env[61006]: value = "task-1337680" [ 1187.239844] env[61006]: _type = "Task" [ 1187.239844] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.247450] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.749653] env[61006]: DEBUG oslo_vmware.api [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Task: {'id': task-1337680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118239} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.750025] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1187.750307] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1187.750627] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1187.750848] env[61006]: INFO nova.compute.manager [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1187.751113] env[61006]: DEBUG oslo.service.loopingcall [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1187.751317] env[61006]: DEBUG nova.compute.manager [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1187.751412] env[61006]: DEBUG nova.network.neutron [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1188.143388] env[61006]: DEBUG nova.compute.manager [req-31020ecd-5831-4e9b-be24-0aa43421c788 req-ba6fdc3c-ecc5-4dab-bbed-8670773f851f service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Received event network-vif-deleted-9b33352b-d3b8-47f4-a5a0-567fa5f8c16a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1188.143597] env[61006]: INFO nova.compute.manager [req-31020ecd-5831-4e9b-be24-0aa43421c788 req-ba6fdc3c-ecc5-4dab-bbed-8670773f851f service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Neutron deleted interface 9b33352b-d3b8-47f4-a5a0-567fa5f8c16a; detaching it from the instance and deleting it from the info cache [ 1188.143765] env[61006]: DEBUG nova.network.neutron [req-31020ecd-5831-4e9b-be24-0aa43421c788 req-ba6fdc3c-ecc5-4dab-bbed-8670773f851f service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.622207] env[61006]: DEBUG nova.network.neutron [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.645693] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27b67b01-3d00-43e6-a2b6-b03bc6d73c91 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.655610] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e343f6-4363-4a4a-b5ca-e5375489d8d4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.677830] env[61006]: DEBUG nova.compute.manager [req-31020ecd-5831-4e9b-be24-0aa43421c788 req-ba6fdc3c-ecc5-4dab-bbed-8670773f851f service nova] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Detach interface failed, port_id=9b33352b-d3b8-47f4-a5a0-567fa5f8c16a, reason: Instance 6ecbaf80-a630-4ffc-949b-69759a2b64a3 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1189.127643] env[61006]: INFO nova.compute.manager [-] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Took 1.38 seconds to deallocate network for instance. [ 1189.634467] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.634748] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.635019] env[61006]: DEBUG nova.objects.instance [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lazy-loading 'resources' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.985404] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.178643] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27ed2d7-cee5-4e5b-8eb3-94b6816ba359 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.186126] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd80e20-de62-47c6-b632-4b564a26ae44 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.215970] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51053446-fef3-4def-b944-f41f688b95de {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.222803] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196a624b-6b2f-4996-ba32-1ccf4da7ee48 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.235549] env[61006]: DEBUG nova.compute.provider_tree [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.459027] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.459221] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Starting heal instance info cache {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1190.459317] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Rebuilding the list of instances to heal {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1190.738457] env[61006]: DEBUG nova.scheduler.client.report [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.977571] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.977714] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquired lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.977861] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Forcefully refreshing network info cache for instance {{(pid=61006) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1190.978019] env[61006]: DEBUG nova.objects.instance [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lazy-loading 'info_cache' on Instance uuid 6ecbaf80-a630-4ffc-949b-69759a2b64a3 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.245171] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.262130] env[61006]: INFO nova.scheduler.client.report [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Deleted allocations for instance 6ecbaf80-a630-4ffc-949b-69759a2b64a3 [ 1191.770197] env[61006]: DEBUG oslo_concurrency.lockutils [None req-a3609747-033e-4d35-8d74-4b8c4980bcc2 tempest-ServerActionsTestOtherB-2061699552 tempest-ServerActionsTestOtherB-2061699552-project-member] Lock "6ecbaf80-a630-4ffc-949b-69759a2b64a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.123s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.983567] env[61006]: DEBUG nova.compute.utils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Can not refresh info_cache because instance was not found {{(pid=61006) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 1192.000549] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Instance cache missing network info. {{(pid=61006) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1192.770477] env[61006]: DEBUG nova.network.neutron [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.273338] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Releasing lock "refresh_cache-6ecbaf80-a630-4ffc-949b-69759a2b64a3" {{(pid=61006) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.273588] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] [instance: 6ecbaf80-a630-4ffc-949b-69759a2b64a3] Updated the network info_cache for instance {{(pid=61006) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1193.273696] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.273867] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.274023] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.777311] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.777679] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.777764] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.777902] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61006) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1193.778812] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34188d43-f101-48ff-ba64-5cd66c732645 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.786708] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e7e6e7-ad1c-4a54-b343-2d1809b68186 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.801264] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712c9541-e5a9-4694-988f-4a6fd0130deb {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.807367] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41b18b0-85ae-4bd7-92ec-6ad59ed738f4 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.835807] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180977MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=61006) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1193.835952] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.836296] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.859005] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Instance bd3aa1cb-f409-416b-9386-448b8bab5b17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61006) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1194.859297] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1194.859362] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61006) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1194.885647] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850382de-01e7-4548-917e-2f913475a2d7 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.893145] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce44054-ff42-4204-988d-258f172ccb2a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.923841] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b909ff07-85de-44d1-81ea-fae23cc3cd3a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.930769] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d147c13-7724-48d5-99c0-75932ac5ff8d {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.943398] env[61006]: DEBUG nova.compute.provider_tree [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.112712] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.112959] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.447354] env[61006]: DEBUG nova.scheduler.client.report [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1195.615960] env[61006]: DEBUG nova.compute.utils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Using /dev/sd instead of None {{(pid=61006) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1195.952148] env[61006]: DEBUG nova.compute.resource_tracker [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61006) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1195.952484] env[61006]: DEBUG oslo_concurrency.lockutils [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.119564] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.137731] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.137960] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.642720] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.642720] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.642720] env[61006]: DEBUG oslo_service.periodic_task [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61006) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.642720] env[61006]: DEBUG nova.compute.manager [None req-76d574b0-fa85-46e1-b05f-055959ee6449 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61006) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1197.192228] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.192615] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.192754] env[61006]: INFO nova.compute.manager [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Attaching volume 0566d828-b0b9-4a9d-ba75-6f02f8279228 to /dev/sdb [ 1197.222722] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f865e8e-73f6-4de8-8646-3bc18b12afc9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.230563] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0155d33e-9e94-4b17-8045-d7efd6160288 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.243467] env[61006]: DEBUG nova.virt.block_device [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating existing volume attachment record: e4def7d0-3f38-4b9f-84bc-16b14a7c546e {{(pid=61006) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1201.789502] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Volume attach. Driver type: vmdk {{(pid=61006) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1201.789762] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285449', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'name': 'volume-0566d828-b0b9-4a9d-ba75-6f02f8279228', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd3aa1cb-f409-416b-9386-448b8bab5b17', 'attached_at': '', 'detached_at': '', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'serial': '0566d828-b0b9-4a9d-ba75-6f02f8279228'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1201.790695] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e60f46-a752-413f-bf6f-bb11b389929a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.806664] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44564ad-5cc8-4df3-8865-24da34a21607 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.829646] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-0566d828-b0b9-4a9d-ba75-6f02f8279228/volume-0566d828-b0b9-4a9d-ba75-6f02f8279228.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.829874] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b254a504-2a12-4118-8613-eb9957ff9033 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.847197] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1201.847197] env[61006]: value = "task-1337684" [ 1201.847197] env[61006]: _type = "Task" [ 1201.847197] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.854110] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.356617] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337684, 'name': ReconfigVM_Task, 'duration_secs': 0.334745} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.356899] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-0566d828-b0b9-4a9d-ba75-6f02f8279228/volume-0566d828-b0b9-4a9d-ba75-6f02f8279228.vmdk or device None with type thin {{(pid=61006) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.361427] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3989033-20a5-42f0-9f0d-29aa78afd532 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.375099] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1202.375099] env[61006]: value = "task-1337685" [ 1202.375099] env[61006]: _type = "Task" [ 1202.375099] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.382487] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.884832] env[61006]: DEBUG oslo_vmware.api [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337685, 'name': ReconfigVM_Task, 'duration_secs': 0.128584} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.885196] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285449', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'name': 'volume-0566d828-b0b9-4a9d-ba75-6f02f8279228', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd3aa1cb-f409-416b-9386-448b8bab5b17', 'attached_at': '', 'detached_at': '', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'serial': '0566d828-b0b9-4a9d-ba75-6f02f8279228'} {{(pid=61006) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1203.920056] env[61006]: DEBUG nova.objects.instance [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid bd3aa1cb-f409-416b-9386-448b8bab5b17 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.424598] env[61006]: DEBUG oslo_concurrency.lockutils [None req-d725e37e-bc22-43ab-9a59-c343da66d4f3 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.232s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.605131] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.605471] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.108690] env[61006]: INFO nova.compute.manager [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Detaching volume 0566d828-b0b9-4a9d-ba75-6f02f8279228 [ 1205.138205] env[61006]: INFO nova.virt.block_device [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Attempting to driver detach volume 0566d828-b0b9-4a9d-ba75-6f02f8279228 from mountpoint /dev/sdb [ 1205.138452] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Volume detach. Driver type: vmdk {{(pid=61006) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1205.138635] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285449', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'name': 'volume-0566d828-b0b9-4a9d-ba75-6f02f8279228', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd3aa1cb-f409-416b-9386-448b8bab5b17', 'attached_at': '', 'detached_at': '', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'serial': '0566d828-b0b9-4a9d-ba75-6f02f8279228'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1205.139513] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bac366-bcfe-4e73-a62b-042faa6e5650 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.161061] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b8f99d-0146-4b72-8a05-f0463a788601 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.167527] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6830a620-d4b3-4278-afe0-c0397d719f92 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.186771] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59eeffcb-1584-450c-bb7d-2dec9d770c76 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.200194] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] The volume has not been displaced from its original location: [datastore2] volume-0566d828-b0b9-4a9d-ba75-6f02f8279228/volume-0566d828-b0b9-4a9d-ba75-6f02f8279228.vmdk. No consolidation needed. {{(pid=61006) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1205.205324] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.205579] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8c5ab62-b8e7-4b46-87f1-1b4da9aef591 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.222752] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1205.222752] env[61006]: value = "task-1337686" [ 1205.222752] env[61006]: _type = "Task" [ 1205.222752] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.229871] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.732148] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337686, 'name': ReconfigVM_Task, 'duration_secs': 0.204652} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.732429] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=61006) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1205.737008] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecaad397-3b46-46c0-8bbc-4840499abde5 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.751241] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1205.751241] env[61006]: value = "task-1337687" [ 1205.751241] env[61006]: _type = "Task" [ 1205.751241] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.758682] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.261106] env[61006]: DEBUG oslo_vmware.api [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337687, 'name': ReconfigVM_Task, 'duration_secs': 0.117849} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.261455] env[61006]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285449', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'name': 'volume-0566d828-b0b9-4a9d-ba75-6f02f8279228', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bd3aa1cb-f409-416b-9386-448b8bab5b17', 'attached_at': '', 'detached_at': '', 'volume_id': '0566d828-b0b9-4a9d-ba75-6f02f8279228', 'serial': '0566d828-b0b9-4a9d-ba75-6f02f8279228'} {{(pid=61006) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1206.803326] env[61006]: DEBUG nova.objects.instance [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'flavor' on Instance uuid bd3aa1cb-f409-416b-9386-448b8bab5b17 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.810980] env[61006]: DEBUG oslo_concurrency.lockutils [None req-bdc50997-f901-4ac2-b8c2-c82b9a919286 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.205s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.837726] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.838161] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.838256] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.838489] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.838672] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.840923] env[61006]: INFO nova.compute.manager [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Terminating instance [ 1208.842718] env[61006]: DEBUG nova.compute.manager [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Start destroying the instance on the hypervisor. {{(pid=61006) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1208.842919] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Destroying instance {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1208.843790] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cab4e9-f6bb-4ce0-b5bd-c363c02213fe {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.851556] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Powering off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1208.851782] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51f9f240-2585-4dfd-9851-29fc5b35cc4a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.857877] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1208.857877] env[61006]: value = "task-1337688" [ 1208.857877] env[61006]: _type = "Task" [ 1208.857877] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.867275] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.367925] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337688, 'name': PowerOffVM_Task, 'duration_secs': 0.170994} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.368206] env[61006]: DEBUG nova.virt.vmwareapi.vm_util [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Powered off the VM {{(pid=61006) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1209.368382] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Unregistering the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1209.368628] env[61006]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e08c8211-1bae-41b6-9b7f-261f0a0d0771 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.431150] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Unregistered the VM {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1209.431366] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Deleting contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1209.431554] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleting the datastore file [datastore2] bd3aa1cb-f409-416b-9386-448b8bab5b17 {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1209.431826] env[61006]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-797932c0-1a8b-4f42-91e2-d6ba34a069aa {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.437566] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for the task: (returnval){ [ 1209.437566] env[61006]: value = "task-1337690" [ 1209.437566] env[61006]: _type = "Task" [ 1209.437566] env[61006]: } to complete. {{(pid=61006) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.444638] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.948388] env[61006]: DEBUG oslo_vmware.api [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Task: {'id': task-1337690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129022} completed successfully. {{(pid=61006) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.948769] env[61006]: DEBUG nova.virt.vmwareapi.ds_util [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted the datastore file {{(pid=61006) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1209.948815] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Deleted contents of the VM from datastore datastore2 {{(pid=61006) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1209.948978] env[61006]: DEBUG nova.virt.vmwareapi.vmops [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Instance destroyed {{(pid=61006) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1209.949177] env[61006]: INFO nova.compute.manager [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1209.949430] env[61006]: DEBUG oslo.service.loopingcall [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61006) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1209.949630] env[61006]: DEBUG nova.compute.manager [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Deallocating network for instance {{(pid=61006) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1209.949726] env[61006]: DEBUG nova.network.neutron [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] deallocate_for_instance() {{(pid=61006) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1210.363378] env[61006]: DEBUG nova.compute.manager [req-1123f0b0-ffde-419e-9870-13ed86990089 req-b781f3c5-e595-4eea-a1d4-03c6b005124e service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Received event network-vif-deleted-7e211d81-36db-43a5-b314-82c59336943a {{(pid=61006) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.363562] env[61006]: INFO nova.compute.manager [req-1123f0b0-ffde-419e-9870-13ed86990089 req-b781f3c5-e595-4eea-a1d4-03c6b005124e service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Neutron deleted interface 7e211d81-36db-43a5-b314-82c59336943a; detaching it from the instance and deleting it from the info cache [ 1210.363737] env[61006]: DEBUG nova.network.neutron [req-1123f0b0-ffde-419e-9870-13ed86990089 req-b781f3c5-e595-4eea-a1d4-03c6b005124e service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.842122] env[61006]: DEBUG nova.network.neutron [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Updating instance_info_cache with network_info: [] {{(pid=61006) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.865962] env[61006]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffe3f9f8-443a-4d33-abbf-5e42cfa7f77a {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.875471] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d9e750-b510-4d52-ae8c-13c7bbbbfd2b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.900084] env[61006]: DEBUG nova.compute.manager [req-1123f0b0-ffde-419e-9870-13ed86990089 req-b781f3c5-e595-4eea-a1d4-03c6b005124e service nova] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Detach interface failed, port_id=7e211d81-36db-43a5-b314-82c59336943a, reason: Instance bd3aa1cb-f409-416b-9386-448b8bab5b17 could not be found. {{(pid=61006) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1211.344744] env[61006]: INFO nova.compute.manager [-] [instance: bd3aa1cb-f409-416b-9386-448b8bab5b17] Took 1.39 seconds to deallocate network for instance. [ 1211.852012] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.852321] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.852568] env[61006]: DEBUG nova.objects.instance [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lazy-loading 'resources' on Instance uuid bd3aa1cb-f409-416b-9386-448b8bab5b17 {{(pid=61006) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1212.387829] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d99ffc-8259-4e42-9002-dc70f4e86e03 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.394720] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b927f619-f39b-480e-be2a-b74f436a5e99 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.424615] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8d6c35-4c9d-4fe6-9061-4c7e73b029e9 {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.430247] env[61006]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5a3b29-0008-4f62-a14f-bd53a449c24b {{(pid=61006) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.442673] env[61006]: DEBUG nova.compute.provider_tree [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed in ProviderTree for provider: 3360045e-46ab-4f2d-9377-dd481ab3cd53 {{(pid=61006) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.945811] env[61006]: DEBUG nova.scheduler.client.report [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Inventory has not changed for provider 3360045e-46ab-4f2d-9377-dd481ab3cd53 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61006) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1213.451317] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.470639] env[61006]: INFO nova.scheduler.client.report [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Deleted allocations for instance bd3aa1cb-f409-416b-9386-448b8bab5b17 [ 1213.978878] env[61006]: DEBUG oslo_concurrency.lockutils [None req-94078b8c-38da-46b7-a946-e3381e17a337 tempest-AttachVolumeNegativeTest-2142602719 tempest-AttachVolumeNegativeTest-2142602719-project-member] Lock "bd3aa1cb-f409-416b-9386-448b8bab5b17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.141s {{(pid=61006) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}